Codementor Events

Hello Microservice Deployment Part 1: Docker

Published Jul 06, 2018Last updated Jan 01, 2019
Hello Microservice Deployment Part 1: Docker

This article is the first of a three part series about deploying microservices.

Before we get started I'd like to make a quick disclaimer: there are affiliate links to books in this series. The books are relevent and they have been very useful to me personally.

In this part, we'll cover the basics of what problems microservices solve (and their causes), create a Docker image, and run it locally as a container.

In Part 2, we'll get our application online by deploying it to a Kubernetes cluster that we set up ourselves on Google Cloud. We'll also deal with the basics of scaling and updating our application.

In Part 3, we'll use Drone.io to set up a simple CI/CD pipeline. It will test our application and roll out any changes we make to the master branch of our repo.

This series is very much the 'hello world' of microservices. We will not cover every aspect of designing, deploying, testing, and monitoring your services (there are entire books written on that topic), but it will give you a pretty good idea of how to get started. I'll also point you to relevent resources along the way so you can deepen your knowledge on your own.

Micro-what?

If you are familiar with the concepts of microservices, feel free to skip this section.

In the beginning, there were monoliths.

Imagine you have an application running on the cloud somewhere that handles: authentication, authorization, sending emails to members, keeping track of fees and membership details, distributing large volumes of entertaining and thought-provoking videos, keeping track of who has seen which video, and making recommendations to members based on their viewing history.

Let's call our totally hypothetical application Webflix.

Let's say we write the entire Webflix application as a single application β€” a monolith. There is quite a lot going on in there, which means a lot of testing would need to happen every time any changes are made, e.g., if I make a change to the code responsible for recommendation, then I should run the entire unit test suite.

After all, there might be some coupling between the recommendation code and some other part of the system. Once my tests pass, I can deploy the code, but that is quite a pain as well because there sure is a lot of it.

I would need to rebuild absolutely everything and deploy a big thing, or introduce some complexity by introducing a kind of packaging system for different code units. Rolling back is similarly inefficient or complex. Once our application is live, there are further inefficiencies.

Let's say we have a few people signed up and paid up, and they are binge watching their respective guilty pleasures one fine Saturday afternoon. With an application like Webflix, one would expect there to be more people watching videos at certain times and on certain days, so it would make sense to build in the capability to scale up and down.

However, when we scale up, we scale everything up since it is a monolith. Wouldn't it be nice to just scale up the parts of the system that are actually under pressure during peak times? For example, the subsystem responsible for actually serving videos should be scaled, but the forgot-password functionality shouldn't.

Enter microservices.

The idea with microservices is that many applications could benefit from being broken down into smaller applications (called services) that play well together. So, going back to Webflix, we would have one service for recommendations, one service for distributing videos, one service for managing payments.

Each service is defined by its purpose and responsibilities. Sometimes, however, things are not so clear. For example, should authentication and authorization be one service or two? Beyond that, Webflix needs to enforce a no-pay-no-watch policy. This means that the authorization system needs to be aware of the payment system. And the payment and forgot-password and user-registration systems all need to be able to send emails.

So we see that sometimes the boundaries between services are pretty clear and obvious. But, sometimes, they can be a little fuzzy. The services also need to communicate effectively.

After breaking things down as above, let's zoom in onto a single well-defined service: the recommendation system. If I make a change to the code of the recommendation service, I just need to run the unit tests for that one system, I can deploy and roll back that service individually and I can scale it individually as needed. I can even deploy multiple versions for a touch of A/B testing.

Winning! But wait, there are other systems involved β€” the recommendation system needs to be accessed by a end-user facing front-end, and it needs to know about what the user has already watched.

That introduces some complexity because there is a chance that a change to the recommendation service will break communication with the functionality that it talks to. There is only so much that can be done with mocks in unit tests. That means that we'll need to build a new layer of testing that makes sure that the various services do in fact play nice.

Then there is the problem of version compatibility. Let's say we are running recommend_service:1.2.0 and video_history_service:1.3.0 and those play nice. But then recommend_service:1.3.0 is created and that one breaks the overall system. So a patch is created to fix it (recommend_service:1.3.1), but, in the meantime, recommend_service:1.4.1 is released.

Alrighty, so let's say we get everything deployed. Now these services need to communicate effectively. We could just use HTTP, but what if the network is just slightly unreliable? What if a service is in the middle of an upgrade? What if the traffic is lost or replayed? Some thought will need to go in there.

What if something breaks without raising an exception? For example, what if "Barney the Dinosaur" is recommended to a hard code zombie fan? The history, user-preference, and recommendations services should all be examined. What if there is a spike in latency? Any number of individual services could be at fault.

There are pros and cons to both microservices and monoliths (monoliths, of course, don't introduce the complexities of microservices β€” communication, deployment, and testing are much easier). The two styles are appropriate for different projects. But, whichever route you go down, you will need to deploy your code. And that is really what this tutorial is about.

Introducing Docker

If you are comfortable with Docker, feel free to skip this section πŸ˜ƒ

deploy-micro-metal.png

You run applications on your computer all the time. Your computer is a bare-metal machine. Applications run on top of your operating system (OS) and the OS manages the hardware. There are a whole host of problems with deploying applications to bare metal, which I wont get into here. Many of those problems are overcome by Virtual Machines(VMs).

deploy-micro-VM.png

A VM runs on a hypervisor, the hypervisor runs on the host OS, and the host OS controls the hardware. There can be multiple unrelated and isolated VMs on top of a single hypervisor. The hypervisor's job is primarily to allocate resources to the various VMs. Now your application will run on top of an OS installed onto a VM.

But the VMs are quite heavy β€” wouldn't it be nice to strip away all of those extra OS's? Containers do that. A container can be thought of as a really lightweight VM.

deploy-micro-containers.png

Containers are much smaller than VMs by default β€” a container doesn't contain a full operating system, whereas a VM does. Containers thus require a lot less in terms of processing power in order to run β€” they tend to be mere megabytes in size and take just seconds to start.

VMs tend to be gigabytes in size and can take minutes to start, because operating systems are big! Containers make use of libraries and packages on the host operating system in order to access resources. They then make use of their own libraries and packages in order to emulate separate operating systems, as needed, instead of installing unnecessary bloat. For example, you could run an Ubuntu image on your Windows machine without needing to install Ubuntu.

VMs (like bare metal) tend to accumulate undocumented bloat as utilities are installed, updated, and generally messed with on the fly. Containers are fully specified in code and easy enough to re-create that manually messing around with their internals usually isn't needed at all.

The smallness of containers is, of course, really great when it comes to scaling applications. If we were hosting recommend_service as a VM and had a spike in traffic, users would just have to deal with some extra latency and perhaps some errors while we brought extra VMs online. On the other hand, if we were using containers, we would be able to bring them online much faster, and potentially even keep extra instances running, in case of spikes, because containers are cheap.

Docker is the de facto industry standard container platform, and that's the one that we'll be dealing with in this article.

Practical: Let's make an API

Docker allows the creation of images. Images are instantiated to create containers (if you are familiar with object orientated programming, then images are like classes, and containers are like objects). In this section, we will create and run a container, and the container will contain a service we wish to deploy.

We'll start off by making a simple Python Hug application and running it locally. Hug is a framework for creating super-fast, self-documenting APIs, no matter how those APIs are exposed.

In our case, we'll be using it to expose a simple API via HTTP. Basically, you create plain old Python functions and then define how (and if) you want them to be exposed through use of decorators. It is fairly new on the scene but considered production ready.

This tutorial assumes you can install Python3, virtualenvwrapper, and Git on your own. You can learn more about Python 3 virtual environments here. Virtualenvwrapper simply provides tooling to make the management of your virtual environments easier.

We're going to start off by running our app locally. We don't actually need to install and run our app locally, but it may add some clarity to the discussion around creating our image.

## create and activate your virtual environment. Application dependencies will be installed here instead of globally. This has nothing to do with containers really, it's just a special directory and path configuration. Also it is good practice

mkvirtualenv --python=`which python3` codementor_deployment_tutorial

## clone the application

git clone https://gitlab.com/sheena.oconnell/tutorial-codementor-deploying-microservices.git

## install the application requirements

cd tutorial-codementor-deploying-microservices
pip install -r requirements.txt

## run the application

hug -f main.py

You should get an output with something like this:

hug -f main.py                                                                                                                                          [11:03]

/#######################################################################\
          `.----``..-------..``.----.
         :/:::::--:---------:--::::://.
        .+::::----##/-/oo+:-##----:::://
        `//::-------/oosoo-------::://.       ###    ###  ###    ###    #####
          .-:------./++o/o-.------::-`   ```  ###    ###  ###    ###  ##
             `----.-./+o+:..----.     `.:///. #########  ###    ### ##
   ```        `----.-::::::------  `.-:::://. ###    ###  ###    ### ###   ####
  ://::--.``` -:``...-----...` `:--::::::-.`  ###    ###  ###   ###   ###    ##
  :/:::::::::-:-     `````      .:::::-.`     ###    ###    #####     ######
   ``.--:::::::.                .:::.`
         ``..::.                .::         EMBRACE THE APIs OF THE FUTURE
             ::-                .:-
             -::`               ::-                   VERSION 2.4.0
             `::-              -::`
              -::-`           -::-
\########################################################################/

 Copyright (C) 2016 Timothy Edmund Crosley
 Under the MIT License


Serving on :8000...
127.0.0.1 - - [07/Jun/2018 11:04:27] "GET /index HTTP/1.1" 200 26

Let's query the index page to make sure the code actually runs. Open a new terminal and then:

curl 0.0.0.0:8000/index

This should return the response:

{"codementor": "so delicious"}

Isn't that nice?

Practical: Let's run our API as a Docker container

Now that you know that the application basically works, it's time to build an image. Notice that the repo contains a Dockerfile. Inside the Dockerfile, you'll find something this:

FROM python:3
EXPOSE 8080

WORKDIR /usr/src/app

COPY requirements.txt ./
RUN pip install --no-cache-dir -r requirements.txt

COPY . .

CMD ["gunicorn", "-b", "0.0.0.0:8080", "main:__hug_wsgi__"]

If you want to learn more about pulling and pushing images, take a look at the docs.

Okay, now we can build the image:

docker build -t codementor-tutorial:1 .

And then to run the image (that is, create a container):

docker run -p 8080:8080 codementor-tutorial:1

You should see output like so:

[2018-06-15 07:11:59 +0000] [1] [INFO] Starting gunicorn 19.8.1
[2018-06-15 07:11:59 +0000] [1] [INFO] Listening at: http://127.0.0.1:8000 (1)
[2018-06-15 07:11:59 +0000] [1] [INFO] Using worker: sync
[2018-06-15 07:11:59 +0000] [9] [INFO] Booting worker with pid: 9

Open up another terminal then curl the API and make sure it all still works:

curl 0.0.0.0:8080/index

This outputs:

{"codementor": "so delicious"}

Notice we are using port 8080 instead of 8000 here. You can specify whatever port you want.

A very brief introduction to versions

So now we have an image that we can run on any computer (bare metal or VM) that can run Docker images. Our image has a name and a version number. If we wanted to make any changes to the functionality of our image, then we would specify those changes in code and then rebuild the image with a different version tag, e.g.:

docker build -t codementor-tutorial:2 .

Summary

Well done πŸ˜ƒ You've managed to build and tag a Docker image and run it as a container.

There is quite a lot more to be said about Docker that is outside the scope of this text. I suggest you take a look at the official documentation if you need more details.

If you need something more structured, there are a lot of truly excellent books out there that can help you along.

Are you ready for the next step? In part 2, we'll be deploying, scaling, and updating our little application on the cloud!

This post contains affiliate links to books that I really enjoy, which means I may receive a commission if you purchase something through these links.

Discover and read more posts from Sheena
get started
post comments7Replies
arielvsp
2 years ago

Nice example. For python 3.9, you’ll need to update the hug version to 2.6.0 in the requirements.txt file. Also in Windows you have to use Admin prompt for most of the installation commands.

Vuyisanani Meteni
3 years ago

Well written and quite informative, Thank you

jia28
6 years ago

Hi,
great information thanks for explaining it!

Show more replies