Skip to content
This repository has been archived by the owner on Aug 17, 2019. It is now read-only.


Switch branches/tags

Name already in use

A tag already exists with the provided branch name. Many Git commands accept both tag and branch names, so creating this branch may cause unexpected behavior. Are you sure you want to create this branch?

Latest commit


Git stats


Failed to load latest commit information.
Latest commit message
Commit time

This repo is archived

The current version of this CSI driver is at

CSI plugin for ScaleIO [Build Status]


CSI-ScaleIO is a Container Storage Interface (CSI) plugin that provides ScaleIO support.

This project may be compiled as a stand-alone binary using Golang that, when run, provides a valid CSI endpoint. This project can also be vendored or built as a Golang plug-in in order to extend the functionality of other programs.

Runtime Dependencies

The Node portion of the plugin can be run on any node that is configured as a ScaleIO SDC. This means that the scini kernel module must be loaded. Also, if the X_CSI_SCALEIO_SDCGUID environment variable is not set, the plugin will try to query the SDC GUID by executing the binary /opt/emc/scaleio/sdc/bin/drv_cfg. If that binary is not present, the Node Service cannot be run.


CSI-ScaleIO can be installed with Go and the following command:

$ go get

The resulting binary will be installed to $GOPATH/bin/csi-scaleio.

If you want to build csi-scaleio with accurate version information, you'll need to run the go generate command and build again:

$ go get
$ cd $GOPATH/src/
$ go generate && go install

The binary will once again be installed to $GOPATH/bin/csi-scaleio.

Start plugin

Before starting the plugin please set the environment variable CSI_ENDPOINT to a valid Go network address such as csi.sock:

$ CSI_ENDPOINT=csi.sock csi-scaleio
INFO[0000] configured com.thecodeteam.scaleio            endpoint="" insecure=true password="******" privatedir=/dev/disk/csi-scaleio sdcGUID= systemname=democluster thickprovision=false user=admin
INFO[0000] identity service registered
INFO[0000] controller service registered
INFO[0000] node service registered
INFO[0000] serving                                       endpoint="unix:///csi.sock"

The server can be shutdown by using Ctrl-C or sending the process any of the standard exit signals.

Using plugin

The CSI specification uses the gRPC protocol for plug-in communication. The easiest way to interact with a CSI plugin is via the Container Storage Client (csc) program provided via the GoCSI project:

$ go get
$ go install

Then, have csc use the same CSI_ENDPOINT, and you can issue commands to the plugin. Some examples...

Get the plugin's supported versions and plugin info:

$ ./csc -e csi.sock identity supported-versions

$ ./csc -v 0.1.0 -e csi.sock identity plugin-info
"com.thecodeteam.scaleio"	"0.0.1+1"
"formed"="Fri, 01 Dec 2017 08:33:28 PST"


When using the plugin, some commands accept additional parameters, some of which may be required for the command to work, or may change the behavior of the command. Those parameters are listed here.

  • CreateVolume: storagepool The name of a storage pool must be passed in the CreateVolume command
  • GetCapacity: storagepool may be passed in GetCapacity command. If it is, the returned capacity is the available capacity for creation within the given storage pool. Otherwise, it's the capacity for creation within the storage cluster.

Passing parameters with csc is demonstrated in this CreateVolume command:

$ ./csc -v 0.1.0 c create --cap 1,mount,xfs --params storagepool=pd1pool1 myvol


The CSI-ScaleIO SP is built using the GoCSI CSP package. Please see its configuration section for a complete list of the environment variables that may be used to configure this SP.

The following table is a list of this SP's default configuration values:

Name Value
X_CSI_PRIVATE_MOUNT_DIR /dev/disk/csi-scaleio

The following table is a list of this configuration values that are specific to ScaleIO, their default values, and whether they are required for operation:

Name Description Default Val Required
X_CSI_SCALEIO_ENDPOINT ScaleIO Gateway HTTP endpoint "" true
X_CSI_SCALEIO_USER Username for authenticating to Gateway "admin" false
X_CSI_SCALEIO_PASSWORD Password of Gateway user "" true
X_CSI_SCALEIO_INSECURE The ScaleIO Gateway's certificate chain and host name should not be verified false false
X_CSI_SCALEIO_SYSTEMNAME The name of the ScaleIO cluster "" true
X_CSI_SCALEIO_SDCGUID The GUID of the SDC. This is only used by the Node Service, and removes a need for calling an external binary to retrieve the GUID "" false
X_CSI_SCALEIO_THICKPROVISIONING Whether to use thick provisioning when creating new volumes false false

Capable operational modes

The CSI spec defines a set of AccessModes that a volume can have. CSI-ScaleIO supports the following modes for volumes that will be mounted as a filesystem:

// Can only be published once as read/write on a single node,
// at any given time.

// Can only be published once as readonly on a single node,
// at any given time.

// Can be published as readonly at multiple nodes simultaneously.

This means that volumes can be mounted to either single node at a time, with read-write or read-only permission, or can be mounted on multiple nodes, but all must be read-only.

For volumes that are used as block devices, only the following are supported:

// Can only be published once as read/write on a single node, at
// any given time.

// Can be published as read/write at multiple nodes
// simultaneously.

This means that giving a workload read-only access to a block device is not supported.

In general, volumes should be formatted with xfs or ext4.


For any questions or concerns please file an issue with the csi-scaleio project or join the Slack channel #project-rexray at


See the VxFlexOS repo for the new driver!







No packages published