fcd2a13 Feb 15, 2018
4 contributors

Users who have contributed to this file

@shlomi-noach @sjmudd @jasonious @derekperkins
129 lines (83 sloc) 6.99 KB

Orchestrator deployment: raft

This text describes deployments for orchestrator on raft.

This complements general deployment documentation.

Backend DB

You may choose between using MySQL and SQLite. See backend configuration.

  • For MySQL:

    • The backend servers will be standalone. No replication setup. Each orchestrator node will interact with its own dedicated backend server.
    • You must have a 1:1 mapping orchestrator:MySQL.
    • Suggestion: run orchestrator and its dedicated MySQL server on same box.
    • Make sure to GRANT privileges for the orchestrator user on each backend node.
  • For SQLite:

    • SQLite is bundled with orchestrator.
    • Make sure the SQLite3DataFile is writable by the orchestrator user.

High availability

orchestrator high availability is gained by using raft. You do not need to take care of backend DB high availability.

What to deploy: service

  • Deploy the orchestrator service onto service boxes. As suggested, you may want to put orchestrator service and MySQL service on same box. If using SQLite there's nothing else to do.

  • Consider adding a proxy on top of the service boxes; the proxy would redirect all traffic to the leader node. There is one and only one leader node, and the status check endpoint is /api/leader-check.

    • Clients may only interact with healthy raft nodes.
  • Nothing should directly interact with a backend DB. Only the leader is capable of coordinating changes to the data with the other raft nodes.

  • orchestrator nodes communicate between themselves on DefaultRaftPort. This port should be open to all orchestrator nodes, and no one else needs to have access to this port.

What to deploy: client

To interact with orchestrator from shell/automation/scripts, you may choose to:

  • Directly interact with the HTTP API

    • You may only interact with the leader. A good way to achieve this is using a proxy.
  • Use the orchestrator-client script.

    • Deploy orchestrator-client on any box from which you wish to interact with orchestrator.

    • Create and edit /etc/profile.d/ on those boxes to read:



      ORCHESTRATOR_API="http://your.orchestrator.service.host1:3000/api http://your.orchestrator.service.host2:3000/api http://your.orchestrator.service.host3:3000/api"

      In the latter case you will provide the list of all orchestrator nodes, and the orchestrator-client script will automatically figure out which is the leader. With this setup your automation will not need a proxy (though you may still wish to use a proxy for web interface users).

      Make sure to chef/puppet/whatever the ORCHESTRATOR_API value such that it adapts to changes in your environment.

  • The orchestrator command line client will refuse to run given a raft setup, since it interacts directly with the underlying database and doesn't participate in the raft consensus, and thus cannot ensure all raft members will get visibility into it changes.

    • Fortunately orchestrator-client provides an almost identical interface as the command line client.
    • You may force the command line client to run via --ignore-raft-setup. This is a "I know what I'm doing" risk you take. If you do choose to use it, then it makes more sense to connect to the leader's backend DB.

Orchestrator service

As noted, a single orchestrator node will assume leadership. Only the leader will:

  • Run recoveries

However all nodes will:

  • Discover (probe) your MySQL topologies
  • Run failure detection
  • Register their own health check

Non-leader nodes must NOT:

  • Run arbitrary commands (e.g. relocate, begin-downtime)
  • Run recoveries per human request.
  • Serve client HTTP requests (but some endpoints, such as load-balancer and health checks, are valid).

A visual example

orchestrator deployment, raft

In the above there are three orchestrator nodes in a raft cluster, each using its own dedicated database (either MySQL or SQLite).

orchestrator nodes communicate with each other.

Only one orchestrator node is the leader.

All orchestrator nodes probe the entire MySQL fleet. Each MySQL server is probed by each of the raft members.

orchestrator/raft operational scenarios

A node crashes:

Start the node, start the MySQL service if applicable, start the orchestrator service. The orchestrator service should join the raft group, get a recent snapshot, catch up with raft replication log and continue as normal.

A new node is provisioned / a node is re-provisioned

Such that the backend database is completely empty/missing.

  • If SQLite, nothing to be done. The node will just join the raft group, get a snapshot from one of the active nodes, catch up with raft log and run as normal.
  • If MySQL, the same will be attempted. However, the MySQL server will have to have the privileges for orchestrator to operate. So if this is a brand new server, those privileges are likely to not be there. As example, our puppet setup periodically ensures privileges are set on our MySQL servers. Thus, when a new server is provisioned, next puppet run lays the privileges for orchestrator. puppet also ensures the orchestrator service is running, and so, pending some time, orchestrator can automatically join the group.
Cloning is valid

If you choose to, you may also provision new boxes by cloning your existing backend databases using your favorite backup/restore or dump/load method.

This is perfectly valid though not required.

Replacing a node

Assuming RaftNodes: ["node1", "node2", "node3"], and you wish to replace node3 with nodeX.

  • You may take down node3, and the raft cluster will continue to work as long as node1 and node2 are good.
  • Create nodeX box. Generate backend db data (see above).
  • On node1, node2 and nodeX reconfigure RaftNodes: ["node1", "node2", "nodeX"].
  • Start orchestrator on nodeX. It will be refused and will not join the cluster because node1 and node2 are not yet aware of the change.
  • Restart orchestrator on node1.
  • Restart orchestrator on node2.
    • All three nodes should form a happy cluster at this time.