Skip to content
Branch: master
Find file Copy path
Find file Copy path
3 contributors

Users who have contributed to this file

@nicolas2bert @RachedBenMustapha @dora-korpar
357 lines (256 sloc) 11.7 KB


Environment Variables



Allows you to run Scality Zenko CloudServer with multiple data backends, defined as regions. When using multiple data backends, a custom locationConfig.json file is mandatory. It will allow you to set custom regions. You will then need to provide associated rest_endpoints for each custom region in your config.json file. Learn more about multiple backends configuration

If you are using Scality RING endpoints, please refer to your customer documentation.

Running it with an AWS S3 hosted backend

To run CloudServer with an S3 AWS backend, you will have to add a new section to your locationConfig.json file with the aws_s3 location type:

"awsbackend": {

"type": "aws_s3", "details": {

"awsEndpoint": "", "bucketName": "yourawss3bucket", "bucketMatch": true, "credentialsProfile": "aws_hosted_profile"




You will also have to edit your AWS credentials file to be able to use your command line tool of choice. This file should mention credentials for all the backends you're using. You can use several profiles when using multiple profiles.

[default] aws_access_key_id=accessKey1 aws_secret_access_key=verySecretKey1 [aws_hosted_profile] aws_access_key_id={{YOUR_ACCESS_KEY}} aws_secret_access_key={{YOUR_SECRET_KEY}}

Just as you need to mount your locationConfig.json, you will need to mount your AWS credentials file at run time: -v ~/.aws/credentials:/root/.aws/credentials on Linux, OS X, or Unix or -v C:\Users\USERNAME\.aws\credential:/root/.aws/credentials on Windows

NOTE: One account can't copy to another account with a source and destination on real AWS unless the account associated with the access Key/secret Key pairs used for the destination bucket has rights to get in the source bucket. ACL's would have to be updated on AWS directly to enable this.



When storing file data, for it to be persistent you must mount docker volumes for both data and metadata. See this section


This is ideal for testing - no data will remain after container is shutdown.


This variable specifies your endpoint. If you have a domain such as, by specifying that here, you and your users can direct s3 server requests to

docker run -d --name s3server -p 8000:8000 -e scality/s3server

Note: In your /etc/hosts file on Linux, OS X, or Unix with root permissions, make sure to associate with


These variables specify authentication credentials for an account named "CustomAccount".

You can set credentials for many accounts by editing conf/authdata.json (see below for further info), but if you just want to specify one set of your own, you can use these environment variables.

docker run -d --name s3server -p 8000:8000 -e SCALITY_ACCESS_KEY_ID=newAccessKey
-e SCALITY_SECRET_ACCESS_KEY=newSecretKey scality/s3server

Note: Anything in the authdata.json file will be ignored. Note: The old ACCESS_KEY and SECRET_KEY environment variables are now deprecated


This variable allows you to change the log level: info, debug or trace. The default is info. Debug will give you more detailed logs and trace will give you the most detailed.

docker run -d --name s3server -p 8000:8000 -e LOG_LEVEL=trace scality/s3server


This variable set to true allows you to run S3 with SSL:

Note1: You also need to specify the ENDPOINT environment variable. Note2: In your /etc/hosts file on Linux, OS X, or Unix with root permissions, make sure to associate with <YOUR_ENDPOINT>

Warning: These certs, being self-signed (and the CA being generated inside the container) will be untrusted by any clients, and could disappear on a container upgrade. That's ok as long as it's for quick testing. Also, best security practice for non-testing would be to use an extra container to do SSL/TLS termination such as haproxy/nginx/stunnel to limit what an exploit on either component could expose, as well as certificates in a mounted volume

docker run -d --name s3server -p 8000:8000 -e SSL=TRUE -e ENDPOINT=<YOUR_ENDPOINT>

More information about how to use S3 server with SSL here


This variable instructs the Zenko CloudServer, and its data and metadata components to listen on the specified address. This allows starting the data or metadata servers as standalone services, for example.

docker run -d --name s3server-data -p 9991:9991 -e LISTEN_ADDR=
scality/s3server npm run start_dataserver


These variables configure the data and metadata servers to use, usually when they are running on another host and only starting the stateless Zenko CloudServer.

docker run -d --name s3server -e DATA_HOST=s3server-data
-e METADATA_HOST=s3server-metadata scality/s3server npm run start_s3server


Use this variable to connect to the redis cache server on another host than localhost.

docker run -d --name s3server -p 8000:8000
-e scality/s3server


Use this variable to connect to the redis cache server on another port than the default 6379.

docker run -d --name s3server -p 8000:8000
-e REDIS_PORT=6379 scality/s3server

Tunables and Setup Tips

Using Docker Volumes

Zenko CloudServer runs with a file backend by default.

So, by default, the data is stored inside your Zenko CloudServer Docker container.

However, if you want your data and metadata to persist, you MUST use Docker volumes to host your data and metadata outside your Zenko CloudServer Docker container. Otherwise, the data and metadata will be destroyed when you erase the container.

docker run -­v $(pwd)/data:/usr/src/app/localData -­v $(pwd)/metadata:/usr/src/app/localMetadata
-p 8000:8000 ­-d scality/s3server

This command mounts the host directory, ./data, into the container at /usr/src/app/localData and the host directory, ./metadata, into the container at /usr/src/app/localMetaData. It can also be any host mount point, like /mnt/data and /mnt/metadata.

Adding modifying or deleting accounts or users credentials

  1. Create locally a customized authdata.json based on our /conf/authdata.json.
  2. Use Docker Volume to override the default authdata.json through a docker file mapping.

For example:

docker run -v $(pwd)/authdata.json:/usr/src/app/conf/authdata.json -p 8000:8000 -d

Specifying your own host name

To specify a host name (e.g., you can provide your own config.json using Docker Volume.

First add a new key-value pair in the restEndpoints section of your config.json. The key in the key-value pair should be the host name you would like to add and the value is the default location_constraint for this endpoint.

For example, is mapped to us-east-1 which is one of the location_constraints listed in your locationConfig.json file here.

More information about location configuration here

"restEndpoints": {
    "localhost": "file",
    "": "file",
    "": "us-east-1"

Then, run your Scality S3 Server using Docker Volume:

docker run -v $(pwd)/config.json:/usr/src/app/config.json -p 8000:8000 -d scality/s3server

Your local config.json file will override the default one through a docker file mapping.

Running as an unprivileged user

Zenko CloudServer runs as root by default.

You can change that by modifing the dockerfile and specifying a user before the entrypoint.

The user needs to exist within the container, and own the folder /usr/src/app for Scality Zenko CloudServer to run properly.

For instance, you can modify these lines in the dockerfile:

&& groupadd -r -g 1001 scality \
&& useradd -u 1001 -g 1001 -d /usr/src/app -r scality \
&& chown -R scality:scality /usr/src/app


USER scality
ENTRYPOINT ["/usr/src/app/"]

Continuous integration with Docker hosted CloudServer

When you start the Docker Scality Zenko CloudServer image, you can adjust the configuration of the Scality Zenko CloudServer instance by passing one or more environment variables on the docker run command line.

Sample ways to run it for CI are:

  • With custom locations (one in-memory, one hosted on AWS), and custom credentials mounted:
docker run --name CloudServer -p 8000:8000
-v $(pwd)/locationConfig.json:/usr/src/app/locationConfig.json
-v $(pwd)/authdata.json:/usr/src/app/conf/authdata.json
-v ~/.aws/credentials:/root/.aws/credentials
-e S3DATA=multiple -e S3BACKEND=mem scality/s3server
  • With custom locations, (one in-memory, one hosted on AWS, one file), and custom credentials set as environment variables (see this section):
docker run --name CloudServer -p 8000:8000
-v $(pwd)/locationConfig.json:/usr/src/app/locationConfig.json
-v ~/.aws/credentials:/root/.aws/credentials
-v $(pwd)/data:/usr/src/app/localData -v $(pwd)/metadata:/usr/src/app/localMetadata
-e S3DATA=multiple -e S3BACKEND=mem scality/s3server

In production with Docker hosted CloudServer

In production, we expect that data will be persistent, that you will use the multiple backends capabilities of Zenko CloudServer, and that you will have a custom endpoint for your local storage, and custom credentials for your local storage:

docker run -d --name CloudServer
-v $(pwd)/data:/usr/src/app/localData -v $(pwd)/metadata:/usr/src/app/localMetadata
-v $(pwd)/locationConfig.json:/usr/src/app/locationConfig.json
-v $(pwd)/authdata.json:/usr/src/app/conf/authdata.json
-v ~/.aws/credentials:/root/.aws/credentials -e S3DATA=multiple
-p 8000:8000 ­-d scality/s3server
You can’t perform that action at this time.