Skip to content
Helm Charts
Smarty Python
Branch: devel
Clone or download
Fetching latest commit…
Cannot retrieve the latest commit at this time.
Type Name Latest commit message Commit time
Failed to load latest commit information.

ESGF Compute Helm Charts




Kubernetes is required to run the ESGF Compute service, installation instructions can be found on the Kubernetes.


Helm is the method of deployment for Kubernetes, installation instructions can be found on their website.


By default this chart will install with persistence disabled and use the latest public images.

There are some configuration files for different environments, use these as references when configuring for your system. These can be combined to setup specific environments.

See the storage section before depoying.

NOTE: There are some required values that need to be set for the chart to install, if these values are missing you may see a similar error to Error: render error in "compute/templates/celery-deployment.yaml": template: compute/templates/celery-deployment.yaml:167:20: executing "compute/templates/celery-deployment.yaml" at <required "Set celery...>: error calling required: Set celery.prometheusUrl. The last portion of the error required: Set celery.prometheusUrl will have which value is missing. In this example the following value needs to be set.

Configuration files
  • development.yaml changes the shared storage and uses development images from a private docker repository.
  • production-ci.yaml enables persistence and uses continuous integration images from a private docker repository.
  • production.yaml enables persistence and uses the latest public images.
  • prometheus-dev.yaml will enable a prometheus instance and provides basic configuration.
  • resources.yaml will define resource requests and limits for each container.
  • traefik-dev.yaml will configure the internal traefik instance to act like the external and internal instance. Without this another intance of traefik must be run.

Following are some examples of combining the above files.

  • helm install . -f production.yaml Setup a production environment.
  • helm install . -f production.yaml -f resources.yaml Setup a production environment with resource definitions.
  • helm install . -f development.yaml -f traefik-dev.yaml -f prometheus-dev.yaml Setup a development environment, configuring traefik as both internal and external load balancer and create a prometheus server instance.
Configuration values
Parameter Description Default
development Disables persistence and health/readiness checks. false
wps.externalHost Sets the external address of the WPS service e.g. Email host address use to send emails from the service. email@host Admin email address, will be cc'd with every registration. admin@host
wps.oauth.client OAuth2 Client ID for ESGF SLCS. fake_oauth_client
wps.oauth.secret OAuth2 Client Secret for ESGF SLCS. fake_oauth_server
wps.apiPassword Password for internal API server. wps_api_password
celery.prometheusUrl Host and path for a prometheus server.

The Helm chart will automatically create all required PersistentVolumes (PV) and PersistentVolumeClaims (PVC). The PVs that are created using HostPath as the storage type. If deploying on a multi-node Kubernetes cluster, the usage of nodeSelector and labeling nodes will be required to ensure persistent storage. The following services use persistent storage in a production environment; Postgres and Redis.


      storage/postgresql: persist
      storage/redis: persist

To disable the creation of the PVs, the path value must be deleted by setting it to null.


  dataClaimName: data-pvc

      storageClassName: slow
      capacity: 100Gi
      path: /p/cscratch/nimbus/wps/data


  dataClaimName: data-pvc

      storageClassName: slow
      capacity: 100Gi
      path: null

Deployment with Tiller

Either choose an existing environment or create a custom one.

  1. git clone
  2. cd docker/helm/compute
  3. helm dependency update
  4. helm install . -f <environement>.yaml will install the chart in the current Kubernetes namespace.
  5. Verify install with helm list and kubectl get pods, see sample.
Deployment without Tiller

Either choose an existing environment or create a custom one.

  1. git clone
  2. cd docker/helm/compute
  3. helm dependency update
  4. helm template . -f production.yaml --output-dir rendered/ will render the templates and create files for Kubernetes.
  5. kubectl apply -k rendered/ will apply all of the files to the current namespace.
  6. Verify install with helm list and kubectl get pods, see sample.

Sample output

The following is sample out from "helm list" and "kubectl get pods"

Helm list output
NAME                    REVISION        UPDATED                         STATUS          CHART           NAMESPACE
precise-alligator       1               Tue Jun 18 00:10:49 2019        DEPLOYED        compute-1.0.0   default
Kubectl get pods
NAME                                                        READY   STATUS    RESTARTS   AGE
precise-alligator-compute-celery-ingress-6f65c48f8c-fgxr9   3/3     Running   0          2m34s
precise-alligator-compute-kube-monitor-8458dd94d5-hxrrw     1/1     Running   0          2m34s
precise-alligator-compute-nginx-7967b9666d-jdnr9            1/1     Running   0          2m34s
precise-alligator-compute-provisioner-7b5c6fdc68-qwlmk      1/1     Running   0          2m34s
precise-alligator-compute-thredds-854864c74d-94zd9          1/1     Running   0          2m34s
precise-alligator-compute-wps-6d7b5dcc66-qbbvc              1/1     Running   0          2m34s
precise-alligator-postgresql-0                              1/1     Running   0          2m34s
precise-alligator-redis-master-0                            1/1     Running   0          2m33s
precise-alligator-traefik-857576cd87-djfhb                  1/1     Running   0          2m34s


In addition to disabling health/readiness checks and persistent storage, some containers can be set to a development mode where their default commands are replaced with /bin/sleep infinity. This allows for live development within the pods container. A directory /devel is created and shared between containers as well. See development.yaml for more information.



In a production environment you may already have an existing Prometheus server. This server will need to be configured to scrap the service endpoints in the Kubernetes cluster.

If you wish to deploy a Prometheus server with Helm there are some included files to assist in this; prometheus.yaml and prometheus-storage.yaml.

celery.prometheusUrl will need to be set to the base address of the prometheus server e.g.

Prometheus Helm Chart

If a development environment is being deployed, the Helm chart has an included Prometheus server. The following command can be used when deploying; helm install . -f development.yaml -f prometheus-dev.yaml. This will deploy the Prometheus chart without any persistent storage. This can be combined with the development Traefik configuration; helm install . -f development.yaml -f traefik-dev.yaml -f prometheus-dev.yaml.


The ESGF Compute service utilizes two instances of Traefik; an external and internal instance. In a production environment traefik.yaml can be used to deploy the external instance. The Helm chart will take care of deploying the internal instance.

In a development environment traefik-dev.yaml can be used to have the internal instance act as both the external and internal instances. When deploying the chart you can using the following command to deploy helm install . -f development.yaml -f traefik-dev.yaml, this will configure the Traefik instance to handle both external and internal Traefik. This can be combined with the development Prometheus server as well; helm install . -f development.yaml -f traefik-dev.yaml -f prometheus-dev.yaml.

You can’t perform that action at this time.