OpsCenter Agents are now preinstalled with all clusters that are OpsCenter enabled. No longer do you have to enter credentials via the OpsCenter GUI.
If you ever need a list of your Cassandra nodes, simply look at /etc/cassandralauncher/nodelist
.
Now you can simply run the DataStax SSH Client by executing: datastax_ssh
. This will connect you to all machines by serially executing your command of choice on all machines in your cluster.
Now you can simply run an ssh c0
or ssh a0
command and easily jump from node to node from within your cluster.
A prefix of c
is reserved for Cassandra nodes, a
for Analytics nodes, and s
for Search nodes. The numerical suffix refers to their order within the ring. View /etc/hosts
on the node for more details.
Now you can simply run the DataStax Parallel SSH Client by executing: datastax_pssh
. This will connect you to all machines and executing your command of choice on all machines in your cluster in parallel.
Now you can use the datastax_s3_store
and datastax_s3_restore
commands to simply upload and download your cassandra/data
directory to and from S3. Think of it as an "experimental EBS functionality for S3."
Note: Do not count on this feature to save important data. This application is only for trivial data in development environments.
The files are stored in s3://datastax_s3_storage-<YOUR_AWS_ACCESS_KEY_ID>/<CLUSTER_NAME>/<NODES_TOKEN>
using the ~/.s3cfg
file.
Using the datastax_pssh
command to store and restore an entire cluster is the recommended way to use this feature.
To automatically preconfigure s3cmd
, which datastax_s3_store
and datastax_s3_restore
rely on, add/change this on your clusterlauncher.conf
:
[S3]
send_s3_credentials = True
That will send .s3cfg preconfigured with your aws_access_key_id and aws_secret_access_key, properly set under 400 permissions, to your home directory.