Skip to content

Latest commit

 

History

History
182 lines (139 loc) · 5.33 KB

6_deploy_on_swarm.md

File metadata and controls

182 lines (139 loc) · 5.33 KB

Deployment on a Docker Swarm

As for the multi Docker host environment, a Docker Swarm requires a key value store to gather the nodes / containers configurations and states.

Creation of a key-value store

Note: if you still have the key-value store from the previous chapter do not re-create it and go directly to the creation of the Swarm.

Several steps are needed to run the key-value store

  • Create dedicated Docker host with Machine) docker-machine create -d virtualbox consul
  • Switch to context of the newly created machine eval "$(docker-machine env consul)"
  • Run container based on Consul image docker run -d -p "8500:8500" -h "consul" progrium/consul -server -bootstrap

Creation of the Swarm

Additional options need to be provided to docker-machine in order to define a Swarm.

Creation of the Swarm master

$ docker-machine create \
-d virtualbox \
--swarm \
--swarm-master \
--swarm-discovery="consul://$(docker-machine ip consul):8500" \
--engine-opt="cluster-store=consul://$(docker-machine ip consul):8500" \
--engine-opt="cluster-advertise=eth1:2376" \
demo0

Creation of the Swarm agent

$ docker-machine create \
 -d virtualbox \
--swarm \
--swarm-discovery="consul://$(docker-machine ip consul):8500" \
--engine-opt="cluster-store=consul://$(docker-machine ip consul):8500" \
--engine-opt="cluster-advertise=eth1:2376" \
demo1

List the nodes

We have created 3 Docker hosts (key-store, Swarm master, Swarm agent)

$ docker-machine ls

NAME     ACTIVE   DRIVER         STATE     URL                         SWARM
consul   *        virtualbox     Running   tcp://192.168.99.100:2376
demo0    -        virtualbox     Running   tcp://192.168.99.101:2376   demo0 (master)
demo1    -        virtualbox     Running   tcp://192.168.99.102:2376   demo1

Create a DNS load balancer

In order to load balance the traffic towards several instances of our app service, we will add a new service. This one uses the DNS round-robin capability of Docker engine (version 1.11) for containers with the same network alias.

Note: to present the DNS round-robin feature, we do not use the load balancer of the previous chapter (dockercloud/haproxy).

The following Dockerfile uses nginx:1.9 official image and add a custom nginx.conf configuration file.

FROM nginx:1.9

# forward request and error logs to docker log collector
RUN ln -sf /dev/stdout /var/log/nginx/access.log
RUN ln -sf /dev/stderr /var/log/nginx/error.log

COPY nginx.conf /etc/nginx/nginx.conf

EXPOSE 80

CMD ["nginx", "-g", "daemon off;"]

The following nginx.conf file define a proxy_pass directive towards http://apps for each request received on port 80.

apps is the value we will set as the app service network alias.

user nginx;
worker_processes 2;
events {
  worker_connections 1024;
}
http {
  access_log /var/log/nginx/access.log;
  error_log /var/log/nginx/error.log;

  # 127.0.0.11 is the address of the Docker embedded DNS server
  resolver 127.0.0.11 valid=1s;
  server {
    listen 80;
    # apps is the name of the network alias in Docker
    set $alias "apps";

    location / {
      proxy_pass http://$alias;
    }
  }
}

Let's build and publish the image of this load-balancer to Docker Cloud:

# Create image
$ docker build -t lucj/lb-dns .

# Publish image
$ docker push -t lucj/lb-dns

The image can now be used in our Docker Compose file.

Update our docker-compose file

The new version of the docker-compose.yml file is the following one

version: '3'
services:
  mongo:
    image: mongo:3.2
    networks:
      - backend
    volumes:
      - mongo-data:/data/db
    expose:
      - "27017"
    environment:
      - "constraint:node==demo0"
  lbapp:
    image: lucj/lb-dns
    networks:
      - backend
    ports:
      - "8000:80"
    environment:
      - "constraint:node==demo0"
  app:
    image: lucj/message-app
    expose:
      - "80"
    environment:
      - MONGO_URL=mongodb://mongo/messageApp
      - "constraint:node==demo1"
    networks:
      backend:
        aliases:
          - apps
    depends_on:
      - lbapp
volumes:
  mongo-data:
networks:
  backend:
    driver: overlay

There are several important updates here

  • usage of the lb-dns image for the load balancer service
  • constraints to choose the nodes on which each service will run (needed in our example to illustrate the DNS round robin)
  • creation of a new user-defined overlay network to enable each container to communicate with each other through their name
  • for each service, definition of the network used
  • definition of network alias for the app service (crucial item as this is the one that will enable nginx to proxy requests)

Deployment and scaling of the application

In order to run the application in this Swarm, we will issue the following commands

  • switch to the swarm master context eval $(docker-machine env --swarm demo0)
  • run the new compose file docker-compose up
  • increase the number of app service instances docker-compose scale app=5

Our application is then available through http://192.168.99.101:8000/message

192.168.99.101 is the IP of the Swarm master. 8000 is the port exported by the load balancer to the outside.