Skip to content
This repository has been archived by the owner on Sep 30, 2022. It is now read-only.

Latest commit

 

History

History
310 lines (252 loc) · 12.7 KB

70_advanced_deployments.rst

File metadata and controls

310 lines (252 loc) · 12.7 KB

Advanced Deployments

Tracking OpenShift objects

As of now, our DeploymentConfigs, Services, Routes etc. are all managed in OpenShift and can only be changed using OpenShift. Wouldn't it be nice if we could track these objects right alongside our code and redeploy not only the application but also all of its configuration automatically? This section will discuss a possible approach for doing exactly what we described.

Disabling ImageStream triggers

Using default configuration, OpenShift will automatically redeploy to our environments if either the DeploymentConfig is modified or a new image is pushed to the ImageStream (using one of our tracked tags). If we would like to update the DeploymentConfig and push a new image in a single step, this would lead to problems (two deployments would be created, the latter of which would probably fail).

This means that we will need to disable deployment triggers and instead only trigger deployments after we have updated the config as well as the image. To disable said deployment triggers, we need to change the triggers property in our DeploymentConfigs to the following:

triggers:
  -
    type: ImageChange
    imageChangeParams:
      automatic: false
      containerNames:
        - webserver-staging
      from:
        kind: ImageStreamTag
        namespace: vshn-demoapp1c
        name: 'webserver:latest'

This is a workaround, as OpenShift doesn't seem to allow completely disabling deployment triggers at the moment.

Triggering deployments from GitLab CI

Having completely disabled deployment triggers, we will need to manually trigger a new deployment every time we push a new image to the APPUiO registry. We will do this by adding an oc deploy command our deployment jobs:

build-staging:
  environment: webserver-staging
  stage: deploy-staging
  image: appuio/gitlab-runner-oc:$OC_VERSION
  services:
    - docker:dind
  script:
    # login to the service account to get access to the internal registry
    - oc login $KUBE_URL --token=$KUBE_TOKEN
    - docker login -u serviceaccount -p `oc whoami -t` $OC_REGISTRY_URL
    # build the docker image and tag it as latest
    # use the current latest image as a caching source
    - docker pull $OC_REGISTRY_IMAGE:latest
    - docker build --cache-from $OC_REGISTRY_IMAGE:latest -t $OC_REGISTRY_IMAGE:latest .
    # push the image to the internal registry
    - docker push $OC_REGISTRY_IMAGE:latest
    # trigger a deployment
    - oc rollout latest dc/webserver-staging
  only:
    - master
  except:
    - tags

The --follow flag in the snippet above allows us to track the progress of the deployment right inside of GitLab CI (it streams the deployment logs).

Exporting configuration objects

In order to add the configuration objects to our repository, we first need to export them from APPUiO. This can be achieved using the oc export command on our staging environment as follows:

$ oc export dc,svc,route webserver-staging
- apiVersion: v1
  kind: DeploymentConfig
  metadata:
    ...
  spec:
    ...
  status:
    ...
- apiVersion: v1
  kind: Service
  metadata:
    ...
  spec:
    ...
  status:
    ...
- apiVersion: v1
  kind: Route
  metadata:
    ...
  spec:
    ...
  status:
    ...

We can now save those configuration objects to separate files in our repository (deployment.yaml, service.yaml, route.yaml). status: and its children can be removed while saving, as this represents the current status of the respective object which is dynamically generated. For the sake of simplicity, we will only track DeploymentConfig, Service and Route in our source control.

Replacing configuration objects using CI

Now that our configuration objects are tracked inside our repository (in docker/openshift/...yaml), we can automatically update the configuration in OpenShift whenever we push a new image. We will then start a new deployment only after the image has been pushed and the new configuration has been updated.

OpenShift allows us to either oc replace an entire configuration object or to oc apply changes to an existing object (which will merge those changes into the existing file). As we track the entire file in our repository and will not want to modify the configuration anywhere but the repository, we will use replace in our approach.

Staging

build-staging:
  environment: webserver-staging
  stage: deploy-staging
  image: appuio/gitlab-runner-oc:$OC_VERSION
  services:
    - docker:dind
  script:
    # login to the service account to get access to the internal registry
    - oc login $KUBE_URL --token=$KUBE_TOKEN
    - docker login -u serviceaccount -p `oc whoami -t` $OC_REGISTRY_URL
    # build the docker image and tag it as latest
    # use the current latest image as a caching source
    - docker pull $OC_REGISTRY_IMAGE:latest
    - docker build --cache-from $OC_REGISTRY_IMAGE:latest -t $OC_REGISTRY_IMAGE:latest .
    # update the configuration in OpenShift
    - oc replace -f docker/openshift -R
    # push the image to the internal registry
    - docker push $OC_REGISTRY_IMAGE:latest
    # trigger a deployment
    - oc rollout latest dc/webserver-staging
  only:
    - master
  except:
    - tags

The oc replace -f docker/openshift -R command will look for configuration objects in our docker/openshift directory and recursively replace all of them on APPUiO. Any changes we might have made using either the CLI or the Web-Interface would be overwritten.

This job will successfully deploy a new configuration and image to the staging environment (as we exported them from the staging environment, their metadata ties them to staging). However, we also want to deploy the exact same configuration to the preprod and prod environment. In order to do this, we will have to dynamically modify their metadata at runtime of the job.

Preprod and prod

To be able to reuse the configuration objects for each environment, we have to dynamically update some metadata. This includes the name of the deployment/service/route as well as the cluster ip of the service.

A simple approach to solving this is the usage of sed as in the snippet below:

variables:
  CLUSTER_IP_STAGING: 172.30.215.173
  CLUSTER_IP_PREPROD: 172.30.29.25
  CLUSTER_IP_PROD: 172.30.31.200
  ...

build-preprod:
  environment: webserver-preprod
  stage: deploy-preprod
  image: appuio/gitlab-runner-oc:$OC_VERSION
  services:
    - docker:dind
  script:
    # login to the service account to get access to the internal registry
    - oc login $KUBE_URL --token=$KUBE_TOKEN
    - docker login -u serviceaccount -p `oc whoami -t` $OC_REGISTRY_URL
    # build the docker image and tag it as stable
    # use the current latest image as a caching source
    - docker pull $OC_REGISTRY_IMAGE:latest
    - docker build --cache-from $OC_REGISTRY_IMAGE:latest -t $OC_REGISTRY_IMAGE:stable .
    # update the configuration in OpenShift
    - sed -i 's;webserver-staging;webserver-preprod;g' docker/openshift/*
    - sed -i 's;webserver:latest;webserver:stable;g' docker/openshift/*
    - sed -i 's;'$CLUSTER_IP_STAGING';'$CLUSTER_IP_PREPROD';g' docker/openshift/*
    - oc replace -f docker/openshift -R
    # push the image to the internal registry
    - docker push $OC_REGISTRY_IMAGE:stable
    # trigger a deployment
    - oc rollout latest dc/webserver-preprod
  only:
    - tags

After we have added those oc replace commands and the necessary sed commands (preprod and prod), our pipelines will automatically deploy configuration alongside the docker image.

Using special YAML features

Currently, our GitLab CI configuration contains quite a bit of duplicate code (even though we have already used variables). YAML allows us to extract duplicate code into a template and include this template in any number of jobs. If we decided to extract duplicates from our test and compile jobs, it would look as follows:

.yarn: &yarn
  stage: build
  image: node:$NODE_VERSION
  cache:
    key: $NODE_VERSION
    paths:
      - node_modules

test:
  <<: *yarn
  script:
    ...

compile:
  <<: *yarn
  script:
    ...

The first line specifies a hidden job and will not be executed by GitLab CI (as it is prefixed by a period). The second part of the first line - &yarn specifically - defines a YAML anchor. This anchor can later be used to refer to this template and include in any number of other jobs.

Lines 11 and 16 are used to refer to our &yarn anchor using <<: *yarn. The <<: directive will merge all children from the template into the test and compile jobs. Everything we have explicitly defined in those jobs will overwrite what is defined in the template (e.g. if we had a script directive in the job as well as in the template)

We can use this behavior to our advantage when implementing templates for the remaining jobs:

.oc: &oc
  image: appuio/gitlab-runner-oc:$OC_VERSION
  script: &oc_script
    # login to the service account to get access to the internal registry
    - oc login $KUBE_URL --token=$KUBE_TOKEN
    - docker login -u serviceaccount -p `oc whoami -t` $OC_REGISTRY_URL
    # build the docker image and tag it as stable
    # use the current latest image as a caching source
    - docker pull $OC_REGISTRY_IMAGE:latest
    - docker build --cache-from $OC_REGISTRY_IMAGE:latest -t $OC_REGISTRY_IMAGE:$DEPLOY_TAG .
    # update the configuration in OpenShift
    - sed -i 's;webserver-staging;webserver-'"$DEPLOY_ENV"';g' docker/openshift/*
    - sed -i 's;webserver:latest;webserver:'"$DEPLOY_TAG"';g' docker/openshift/*
    - sed -i 's;'$CLUSTER_IP_STAGING';'$CLUSTER_IP';g' docker/openshift/*
    - oc replace -f docker/openshift -R
    # push the image to the internal registry
    - docker push $OC_REGISTRY_IMAGE:$DEPLOY_TAG
    # trigger a deployment
    - oc rollout latest dc/webserver-$DEPLOY_ENV

build-staging:
  <<: *oc
  environment: webserver-staging
  stage: deploy-staging
  services:
    - docker:dind
  only:
    - master
  except:
    - tags
  variables:
    CLUSTER_IP: 172.30.215.173
    DEPLOY_ENV: staging
    DEPLOY_TAG: latest

build-preprod:
  <<: *oc
  environment: webserver-preprod
  stage: deploy-preprod
  services:
    - docker:dind
  only:
    - tags
  variables:
    CLUSTER_IP: 172.30.29.25
    DEPLOY_ENV: preprod
    DEPLOY_TAG: stable

build-prod:
  <<: *oc
  environment: webserver-prod
  stage: deploy-prod
  script:
    # execute prod specific scripts
    ...
  only:
    - tags
  when: manual
  variables:
    CLUSTER_IP: 172.30.31.200

In the above configuration, as much as possible has been extracted into the hidden job .oc which is included in all deployment jobs (see lines 22, 37 and 50). To make environment specific values available to the template, CLUSTER_IP, DEPLOY_ENV and DEPLOY_TAG have been extracted into job specific variables (as in lines 10, 12-14, 17, 19 and 32-34).

As the last job doesn't build a new docker image, its script differs from the other two deployment jobs. If we explicitly specify a script while also including from a template, the explicitly defined script will always take precedence (see lines 53-55).