Zenoss Deployment Tool
The primary purpose of the playbook is to deploy Zenoss Core to your own bare bone server or a virtual host:
- deploy docker with lvm storage thin pool
- deploy serviced according to requirements optionally deploy a serviced cluster, which in the turn can be used to implement zenoss distributed monitoring
- deploy and run Zenoss Core
- Support Zenoss deployments to Debian and Ubuntu operation systems.
- Fully automated deployments
- Cloud ready: one click Google Cloud Engine deployment
Install Ansible. Put your host to the hosts and configure lvm_dev variable. Proceed to Usage section :)
- lvm_dev : block device for lvm and docker thin pool. the value of global variable will be used if not set.
- pool_name : name of serviced pool to assign host to. the default pool is used if not set.
- prepare : prepare environment
- docker : deploy docker
- serviced : deploy serviced
- zenoss : deploy zenoss
After required configuration prepared you can use setup script to start deployment.
At least one spare partition(>= 30 Gb) must be available and configured(lvm_dev) for docker and zenoss storage pools. 4 Gb RAM to run serviced services or 24 Gb to run zenoss
- Redhat / CentOS 7
- Debian Stretch
- Ubuntu Xenial, Bionic
ansible: >= 2.3.0
Google Cloud Deployment
The playbook tested for serviced 1.5.1 and zenoss 6.2.1. Google cloud host is used for deployment as a reproducible clean environment. You have to setup Terraform than zenoss.tf will be used to create GCE instance.
Run gce_deploy script for full cycle deployment to google cloud:
./gce_deploy <your google cloud ssh key username> [ terraform options ]