Skip to content
Branch: master
Find file History
bitnami-bot [bitnami/node] Release 10.0.4 updating components versions
Signed-off-by: Bitnami Containers <>
Latest commit 4c021a8 Aug 16, 2019


Node Event-driven I/O server-side JavaScript environment based on V8.


$ helm repo add bitnami
$ helm install bitnami/node


This chart bootstraps a Node deployment on a Kubernetes cluster using the Helm package manager.

It clones and deploys a Node.js application from a Git repository. Optionally, you can set up an Ingress resource to access your application and provision an external database using the Kubernetes service catalog and the Open Service Broker for Azure.

Bitnami charts can be used with Kubeapps for deployment and management of Helm Charts in clusters. This Helm chart has been tested on top of Bitnami Kubernetes Production Runtime (BKPR). Deploy BKPR to get automated TLS certificates, logging and monitoring for your applications.


  • Kubernetes 1.4+ with Beta APIs enabled
  • PV provisioner support in the underlying infrastructure

Installing the Chart

To install the chart with the release name my-release:

$ helm repo add bitnami
$ helm install --name my-release bitnami/node

These commands deploy Node.js on the Kubernetes cluster in the default configuration. The configuration section lists the parameters that can be configured during installation. Also includes support for MariaDB chart out of the box.

Due that the Helm Chart clones the application on the /app volume while the container is initializing, a persistent volume is not required.

Tip: List all releases using helm list

Uninstalling the Chart

To uninstall/delete the my-release deployment:

$ helm delete my-release

The command removes all the Kubernetes components associated with the chart and deletes the release.


The following table lists the configurable parameters of the Node chart and their default values.

Parameter Description Default
global.imageRegistry Global Docker image registry nil
global.imagePullSecrets Global Docker registry secret names as an array [] (does not add image pull secrets to deployed pods)
image.registry NodeJS image registry
image.repository NodeJS image name bitnami/node
image.tag NodeJS image tag {TAG_NAME}
image.pullPolicy NodeJS image pull policy IfNotPresent
image.pullSecrets Specify docker-registry secret names as an array [] (does not add image pull secrets to deployed pods)
nameOverride String to partially override node.fullname template with a string (will prepend the release name) nil
fullnameOverride String to fully override node.fullname template with a string nil
volumePermissions.enabled Enable init container that changes volume permissions in the data directory (for cases where the default k8s runAsUser and fsUser values do not work) false
volumePermissions.image.registry Init container volume-permissions image registry
volumePermissions.image.repository Init container volume-permissions image name bitnami/minideb
volumePermissions.image.tag Init container volume-permissions image tag latest
volumePermissions.image.pullPolicy Init container volume-permissions image pull policy Always
volumePermissions.resources Init container resource requests/limit nil
git.registry Git image registry
git.repository Git image name bitnami/git
git.tag Git image tag {TAG_NAME}
git.pullPolicy Git image pull policy IfNotPresent
repository Repo of the application
revision Revision to checkout master
replicas Number of replicas for the application 1
applicationPort Port where the application will be running 3000
extraEnv Any extra environment variables to be pass to the pods {}
securityContext.enabled Enable security context true
securityContext.fsGroup Group ID for the container 1001
securityContext.runAsUser User ID for the container 1001
service.type Kubernetes Service type ClusterIP
service.port Kubernetes Service port 80
service.annotations Annotations for the Service {}
service.loadBalancerIP LoadBalancer IP if Service type is LoadBalancer nil
service.nodePort NodePort if Service type is LoadBalancer or NodePort nil
persistence.enabled Enable persistence using PVC false
persistence.path Path to persisted directory /app/data
persistence.accessMode PVC Access Mode ReadWriteOnce
persistence.size PVC Storage Request 1Gi
mongodb.install Wheter to install or not the MongoDB chart true
externaldb.secretName Secret containing existing database credentials nil
externaldb.type Type of database that defines the database secret mapping osba The existing ServiceInstance to be used nil
ingress.enabled Enable ingress controller resource false
ingress.hosts[0].name Hostname to your Node installation node.local
ingress.hosts[0].path Path within the url structure /
ingress.hosts[0].tls Utilize TLS backend in ingress false
ingress.hosts[0].certManager Add annotations for cert-manager false
ingress.hosts[0].tlsSecret TLS Secret (certificates) node.local-tls-secret
ingress.hosts[0].annotations Annotations for this host's ingress record []
ingress.secrets[0].name TLS Secret Name nil
ingress.secrets[0].certificate TLS Secret Certificate nil
ingress.secrets[0].key TLS Secret Key nil

The above parameters map to the env variables defined in bitnami/node. For more information please refer to the bitnami/node image documentation.

Specify each parameter using the --set key=value[,key=value] argument to helm install. For example,

$ helm install --name my-release \
  --set repository=,replicas=2 \

The above command clones the remote git repository to the /app/ directory of the container. Additionally it sets the number of replicas to 2.

Alternatively, a YAML file that specifies the values for the above parameters can be provided while installing the chart. For example,

$ helm install --name my-release -f values.yaml bitnami/node

Tip: You can use the default values.yaml

Rolling VS Immutable tags

It is strongly recommended to use immutable tags in a production environment. This ensures your deployment does not change automatically if the same tag is updated with a different image.

Bitnami will release a new chart updating its containers if a new version of the main container, significant changes, or critical vulnerabilities exist.


The Bitnami Node image stores the Node application and configurations at the /app path of the container.

Persistent Volume Claims are used to keep the data across deployments. This is known to work in GCE, AWS, and minikube. See the Configuration section to configure the PVC or to disable persistence.

Adjust permissions of persistent volume mountpoint

As the image run as non-root by default, it is necessary to adjust the ownership of the persistent volume so that the container can write data into it.

By default, the chart is configured to use Kubernetes Security Context to automatically change the ownership of the volume. However, this feature does not work in all Kubernetes distributions. As an alternative, this chart supports using an initContainer to change the ownership of the volume before mounting it in the final destination.

You can enable this initContainer by setting volumePermissions.enabled to true.

Set up an Ingress controller

First install the nginx-ingress controller via helm:

$ helm install stable/nginx-ingress

Now deploy the node helm chart:

$ helm install --name my-release bitnami/node --set ingress.enabled=true,,service.type=ClusterIP

Configure TLS termination for your ingress controller

You must manually create a secret containing the certificate and key for your domain. You can do it with this command:

$ kubectl create secret tls my-tls-secret --cert=path/to/file.cert --key=path/to/file.key

Then ensure you deploy the Helm chart with the following ingress configuration:

  enabled: false
  path: /
  annotations: nginx

Connect your application to an already existing database

  1. Create a secret containing your database credentials:
$ kubectl create secret generic my-database-secret --from-literal=host=YOUR_DATABASE_HOST --from-literal=port=YOUR_DATABASE_PORT --from-literal=username=YOUR_DATABASE_USER  --from-literal=password=YOUR_DATABASE_PASSWORD --from-literal=database=YOUR_DATABASE_NAME


  1. Deploy the node chart specifying the secret name
$ helm install --name node-app --set mongodb.install=false,externaldb.secretName=my-database-secret bitnami/node

Provision a database using the Open Service Broker for Azure

  1. Install Service Catalog in your Kubernetes cluster following this instructions
  2. Install the Open Service Broker for Azure in your Kubernetes cluster following this instructions

TIP: you may want to install the osba chart setting the modules.minStability=EXPERIMENTAL to see all the available services.

$ helm install azure/open-service-broker-azure --name osba --namespace osba \
       --set azure.subscriptionId=$AZURE_SUBSCRIPTION_ID \
       --set azure.tenantId=$AZURE_TENANT_ID \
       --set azure.clientId=$AZURE_CLIENT_ID \
       --set azure.clientSecret=$AZURE_CLIENT_SECRET \
       --set modules.minStability=EXPERIMENTAL
  1. Create and deploy a ServiceInstance to provision a database server in Azure cloud.
kind: ServiceInstance
  name: azure-mongodb-instance
    app: mongodb
  clusterServiceClassExternalName: azure-cosmosdb-mongo-account
  clusterServicePlanExternalName: account
    resourceGroup: mongodb-k8s-service-catalog
      -  ""

Please update the YOUR_AZURE_LOCATION placeholder in the above example.

$ kubectl create -f mongodb-service-instance.yml
  1. Deploy the helm chart:

    $ helm install --name node-app --set mongodb.install=false,,externaldb.ssl=true bitnami/node

Once the instance has been provisioned in Azure, a new secret should have been automatically created with the connection parameters for your application.

Deploying the helm chart enabling the Azure external database makes the following assumptions:

  • You would want an Azure CosmosDB MongoDB database
  • Your application uses DATABASE_HOST, DATABASE_PORT, DATABASE_USER, DATABASE_PASSWORD, and DATABASE_NAME environment variables to connect to the database.

You can read more about the kubernetes service catalog at

Notable changes


This release includes security contexts, so the containers in the chart are run as non-root. More information in this link.


To 6.0.0

Backwards compatibility is not guaranteed unless you modify the labels used on the chart's deployments. Use the workaround below to upgrade from versions previous to 6.0.0. The following example assumes that the release name is node:

$ kubectl patch deployment node --type=json -p='[{"op": "remove", "path": "/spec/selector/matchLabels/chart"}]'
$ kubectl patch deployment node-mongodb --type=json -p='[{"op": "remove", "path": "/spec/selector/matchLabels/chart"}]'
You can’t perform that action at this time.