In the previous article "Automate MERN App Deployment with GitHub Actions CI/CD", we learned till pushing images to the docker hub. So far, you've done an excellent job! But that isn't enough; we now require a server that runs our containers 24 hours a day, seven days a week, and should be accessible to anyone with an internet connection. It is not a good idea to set up your own server ๐ . So we'll need to rent a server. Who will give us a rented server and how to set it up are what you will learn in this article.
Let's Get Started
Prerequisites
Get source code from here if haven't followed previous tutorials in this series.
Read previous articles in this series so that you won't get confused.
All the AWS services we use in this tutorial are free but you need a credit/debit card to create an account.
Before we proceed, note that our super simple productivity-app doesn't need AWS. This is just for learning purposes.
โ ๏ธ Warning: Don't forget to turn off any running instances, and environments as soon as you finish this tutorial to avoid getting billed.
What is AWS?
As I mentioned in the beginning, we need someone who will let us rent a server and that is what AWS is about. Not physically like a room and bunch of boxes but on the cloud means you can rent servers from anywhere in the world.
This is how it looks like if you set up your own servers physically ๐:
If you haven't been living under a rock, you've probably heard of AWS. It is an abbreviation for Amazon Web Services. AWS (by Amazon) offers a wide range of cloud computing services, such as computing, storage, database, analytics, machine learning, networking, mobile, developer tools, security, and enterprise applications. for all of your requirements, eliminating the need for you to set up your own servers. It provides a flexible and scalable infrastructure that can be tailored to each user's unique requirements, and it is widely regarded as one of the leading cloud platforms available today. Google Cloud (by Google) and Azure (by Microsoft) are both major competitors to AWS.
Big Brain Time ๐ก: If you are following from the beginning, in the first article "Let's build and deploy a full stack MERN web application", we deployed our frontend application on Netlify and backend on Heroku. Those are fine but most companies need IaaS (Infrastructure as a Server) over PaaS (Platform as a Server).
Infrastructure as a Service (IaaS) VS Platform as a Service (PaaS)
This image clearly shows the difference between Iaas and PaaS:
Setting up NGINX
NGINX is a popular web server that is frequently used as a load balancer, reverse proxy, and HTTP cache. Consider it a traffic controller that manages traffic across multiple servers based on how busy a server is.
As a reverse proxy, NGINX serves as a go-between for the client and the server. When you (the client) send a request to the server, it is routed through the reverse proxy and then to the web server. This reverse proxy is useful in several ways: it can distribute requests across multiple servers, cache frequently-requested content, reduce response time, and protect against malicious requests.
If you want to learn more about NGINX, I would recommend reading "The NGINX Handbook" written by Farhan Hasin Chowdhury on freeCodeCamp.
Okay! Why do we care? Instead of deploying frontend and backend separately, we can simply use a load balancer like NGINX to combine both the frontend and backend into one web server, which can then distribute incoming requests to the appropriate server based on the type of request and server load. This simplifies the deployment process and helps to ensure better performance and scalability of the application.
Configuring NGINX
Open the project in your code editor. In the root, create a folder called nginx
.
In that folder, create a file called nginx.conf
and paste this code.
# Defining a server group called `client` that has one server in it called `client` on port `3000`.
upstream client {
server client:3000;
}
# Defining a server group called `server` that has one server in it called `server` on port `5000`.
upstream server {
server server:5000;
}
# Listening on port 80 and redirecting requests to the client and server.
server {
listen 80;
listen [::]:80;
proxy_http_version 1.1;
proxy_set_header Upgrade $http_upgrade;
proxy_set_header Connection 'upgrade';
proxy_set_header Host $host;
proxy_cache_bypass $http_upgrade;
location / {
proxy_pass http://client;
}
location /server {
rewrite /server/(.*) /$1 break;
proxy_pass http://server;
}
}
The
upstream
directive is used to define the server groups, one for theclient
on port3000
and another for theserver
on port5000
.The
server
directive is used to define the server that listens on port80
for incoming requests. Thelisten
directive specifies the port and theproxy_http_version
,proxy_set_header
, andproxy_cache_bypass
directives are used to configure the proxy settings.The
location
directive is used to define the URL paths that are redirected to each server group. The firstlocation /
directive redirects all requests to theclient
server group. The secondlocation /server
directive matches any request with the path/server
and uses therewrite
directive to remove/server
from the path before redirecting to theserver
server group using theproxy_pass
directive.
Even I'm using NGINX for the first time, so please read the article I mentioned above to clearly understand how to write NGINX config files.
Now, create Dockerfile in that same folder and paste this.
FROM nginx:stable-alpine
RUN rm /etc/nginx/conf.d/*
COPY ./nginx.conf /etc/nginx/conf.d/
CMD [ "nginx", "-g", "daemon off;" ]
Setting up AWS
We are going to use three AWS services:
Elastic Beanstalk - A fully-managed service that makes it easy to deploy applications.
EC2 - Elastic Compute is a computing service that offers scalable computing capacity. You can quickly start a virtual machine with your preferred configurations such as operating system, software, storage, and more.
S3 Bucket - A highly scalable storage service. All of your app data will be stored here.
Only Elastic Beanstalk needs to be configured; the other two will be configured automatically.
So, go ahead and log in to your AWS management console.
Create a user
To begin, we must create a user who will manage cloud resources. Because we're automating the deployment with GitHub Actions, we need a way for Actions to modify our resources, such as deploying code to AWS, and that's what this user is for. Make sure to grant the appropriate permissions.
In the search bar, type "IAM" (Identity and Access Management) and click on Users under Access management.
Click on Add users and give your user a name -> click Next.
In Set permissions select Attach policies directly -> check AdministratorAccess-AWSElasticBeanstalk -> click Next -> finally click Create user.
Create and configure an application in Elastic Beanstalk
Again in the search bar, type "Elastic Beanstalk".
Click Create Application.
Give your application a name, leave Application tags blank, select "Docker" as Platform, and leave the rest default. Click Create application.
Congrats ๐! Your application has been created. You can visit it by clicking the URL shown in the application environment. This is a sample application provided by EB.
Configure environment variables
Before we deploy our application we need to add environment variables in EB.
To do so, go to the application Configuration page and click Edit under the Software category.
Scroll to the very bottom of the page -> under Environment variables and add the following env variables.
REACT_APP_BACKEND_URL = /server
MONGODB_URI = your MongoDB URI
TOKEN_KEY = random string
EMAIL = email
PASSWORD = password
That's it! It's now time to add the deployment step to our CI/CD pipeline.
Deploying MERN App on AWS with CI/CD
Open pipeline.yml
and add this step right below pushing images to the docker hub step.
# This is the step that is deploying the application to Elastic Beanstalk.
- name: Deploy to Elastic Beanstalk
uses: einaregilsson/beanstalk-deploy@v21
with:
aws_access_key: ${{ secrets.AWS_ACCESS_KEY_ID }}
aws_secret_key: ${{ secrets.AWS_SECRET_ACCESS_KEY }}
application_name: ${{ secrets.EB_APP_NAME }}
environment_name: ${{ secrets.EB_ENV_NAME }}
region: ${{ secrets.EB_REGION }}
version_label: "version-${{ github.run_number }}"
deployment_package: docker-compose.yml
Here we are using a pre-defined GitHub action
einaregilsson/beanstalk-deploy@v21
so that we don't need to write from scratch. There's nothing needed to explain, it's pretty straightforward.Add
AWS_ACCESS_KEY_ID
,AWS_SECRET_ACCESS_KEY
to GitHub secrets. You can get these from the IAM dashboard. Just click on the user you created and go to the Security credentials tab and copy from there.EB_APP_NAME
is your EB application name andEB_ENV_NAME
is your application environment name (ends with "-env" in my case it is "Productivityapp-env").EB_REGION
is the region assigned to you.
Because docker-compose cannot read GitHub environment variables, we are now creating a second image with the word 'latest' at the end of the image tag instead of github.run
_number
. This is the final YAML file. (You can build a single image if you want, just remove the tag with github.run
_number
)
# The name of the workflow.
name: Build and Deploy
# Run the workflow when code is pushed to the main branch
on:
push:
branches:
- main
# Set environment variables
env:
MONGODB_URI: ${{ secrets.MONGODB_URI }}
TOKEN_KEY: ${{ secrets.TOKEN_KEY }}
EMAIL: ${{ secrets.EMAIL }}
PASSWORD: ${{ secrets.PASSWORD }}
# This is the workflow that is being run.
jobs:
build-and-deploy:
# This is telling GitHub to run the workflow on the latest version of Ubuntu.
runs-on: ubuntu-latest
steps:
# Checkout the code from the GitHub repository
- name: Checkout code
uses: actions/checkout@v3
# Install dependencies and run tests for the client application
- name: Install and Test Client
working-directory: ./client
run: |
npm install
npm run test
# Install dependencies, export environment variables to be used by application and run tests for the server application
- name: Install and Test Server
working-directory: ./server
run: |
npm install
export MONGODB_URI=$MONGODB_URI
export TOKEN_KEY=$TOKEN_KEY
export EMAIL=$EMAIL
export PASSWORD=$PASSWORD
npm run test
# Build a Docker image for the client application
- name: Build Client Docker Image
working-directory: ./client
# Build image with tag rakeshpotnuru/productivity-app:client
run: |
docker build -t rakeshpotnuru/productivity-app:client-${{github.run_number}} -t rakeshpotnuru/productivity-app:client-latest .
# Build a Docker image for the server application
- name: Build Server Docker Image
working-directory:
./server
# Build image with tag rakeshpotnuru/productivity-app:server
run: |
docker build -t rakeshpotnuru/productivity-app:server-${{github.run_number}} -t rakeshpotnuru/productivity-app:server-latest .
# Build a Docker image for the NGINX reverse proxy
- name: Build NGINX Docker Image
working-directory: ./nginx
# Build image with tag rakeshpotnuru/productivity-app:nginx
run: |
docker build -t rakeshpotnuru/productivity-app:nginx-${{github.run_number}} -t rakeshpotnuru/productivity-app:nginx-latest .
# Login to Docker Hub using credentials from repository secrets
- name: Log in to Docker Hub
uses: docker/login-action@v2
with:
username: ${{ secrets.DOCKER_USERNAME }}
password: ${{ secrets.DOCKER_PASSWORD }}
# Push the Docker images to Docker Hub
- name: Push Docker Images to Docker Hub
run: |
docker push rakeshpotnuru/productivity-app:client-${{github.run_number}}
docker push rakeshpotnuru/productivity-app:server-${{github.run_number}}
docker push rakeshpotnuru/productivity-app:nginx-${{github.run_number}}
docker push rakeshpotnuru/productivity-app:client-latest
docker push rakeshpotnuru/productivity-app:server-latest
docker push rakeshpotnuru/productivity-app:nginx-latest
# This is the step that is deploying the application to Elastic Beanstalk.
- name: Deploy to Elastic Beanstalk
uses: einaregilsson/beanstalk-deploy@v21
with:
aws_access_key: ${{ secrets.AWS_ACCESS_KEY_ID }}
aws_secret_key: ${{ secrets.AWS_SECRET_ACCESS_KEY }}
application_name: ${{ secrets.EB_APP_NAME }}
environment_name: ${{ secrets.EB_ENV_NAME }}
region: ${{ secrets.EB_REGION }}
version_label: "version-${{ github.run_number }}"
deployment_package: docker-compose.yml
deployment_package
There are two ways you can provide a package for deployment package - 1. Zip file or 2. docker-compose file. Since we are pushing images to the docker hub we will use the docker-compose.yml
file.
Rename the previous docker-compose.yml
file as docker-compose.dev
.yml
and create a new docker-compose.yml
file. Paste this code.
version: "3.8"
services:
nginx:
image: rakeshpotnuru/productivity-app:nginx-latest
restart: always
depends_on:
- client
- server
ports:
- "80:80"
client:
image: rakeshpotnuru/productivity-app:client-latest
environment:
- CHOKIDAR_USEPOLLING=true
- REACT_APP_BACKEND_URL=${REACT_APP_BACKEND_URL}
server:
image: rakeshpotnuru/productivity-app:server-latest
environment:
- MONGODB_URI=${MONGODB_URI}
- TOKEN_KEY=${TOKEN_KEY}
- EMAIL=${EMAIL}
- PASSWORD=${PASSWORD}
- Instead of re-building Docker images with Dockerfile, we pull pre-built images from Docker Hub. And those environment variables are derived from variables we previously defined in the EB application environment.
That's all! Push the code to GitHub and allow all checks in the GitHub Actions workflow to pass. You can see your application in action ๐๐๐คฉ.
Terminating the Environment
Don't forget to terminate the environment after finishing this tutorial if you no longer use that application.
To terminate, go to Environments -> select the environment you want to terminate -> click Actions -> click Terminate environment. This will also terminate any instances that were created by your application.
If you've made it this far from "Let's build and deploy a full stack MERN web application", kudos. Give yourself a shoulder pat.
This is not the end of the story. Many exciting articles are on the way! Follow me for more - Onwards and Upwards ๐!