-
Notifications
You must be signed in to change notification settings - Fork 7
/
cluster.ini
101 lines (76 loc) · 2.7 KB
/
cluster.ini
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
[network]
# Consul, Vault and Nomad will be bound to this address.
# This field is required.
address = 10.66.60.1
# Nomad will bind services on this interface.
# This field is required.
interface = liquid-bridge
# Create a local bridge with the name and address listed above.
# See scripts/create-bridge.sh
create_bridge = true
# Set iptables DNAT rules to forward PUBLIC:PRIVATE port pairs.
# Uses the `forward_address` below.
# See scripts/forward-ports.sh
forward_ports = 80:80,443:443
# Which interface should the DNAT listen on? Auto-detect first public IP
# address if blank.
#forward_address = 192.168.0.2
[nomad]
; version = 0.9.0
# Time until dead jobs are no longer displayed in the UI.
; zombie_time = 4h
# Drain node on "./cluster.py stop" and "docker stop".
# When restarting the node, if the data has not been deleted,
# the services will re-appear.
# Set this to `false` to attempt zero-downtime upgrades.
; drain_on_stop = false
# Delete ./var/nomad (allocations, images and past jobs) on boot.
# Enable this if Nomad boots up stray jobs after
# an upgrade. Having this enabled means you need to start all jobs
# on every restart.
; delete_data_on_start = true
# Number of scheduler threads for this nomad worker.
# Defaults to 2.
; nomad_schedulers = 2
# Percentage of physical memory to use by nomad. Valid values: [10-90]. Default value: 70.
; memory_percent = 70
# Set user-defined meta flags for this Nomad node.
[nomad_meta]
; storage = true
; meaning_of_life = happiness
# These flags are used by the Liquid Investigations CI.
volumes = /opt/volumes
cluster_volumes = /opt/cluster/var/volumes
liquid_volumes = /opt/node/volumes
liquid_collections = /opt/node/collections
liquid_ingress = true
[vault]
; version = 1.1.1
# If vault complains with `Failed to lock memory`, either allow the vault
# binary to call mlock (`sudo setcap cap_ipc_lock=+ep bin/vault`), or
# disable mlock after reading
# https://www.vaultproject.io/docs/configuration/#disable_mlock
disable_mlock = false
[consul]
; version = 1.4.4
[cluster]
# Name of this Nomad and Consul node. Defaults to the machine hostname.
; node_name = liquid-cluster
# Run Consul, Vault and Nomad in -dev mode.
dev = false
# Set the log level to DEBUG.
debug = false
# Set to "true" to run this node in client-only mode.
client = false
# Multi-host configuration
; bootstrap_expect = 3
; retry_join = 10.66.60.1,10.66.60.2,10.66.60.4
# Configure higher memory limits for services. Use this in case you encounter
# "OOM Killed" messages. Start with these higher limits:
; influxdb_memory_limit = 1024
; prometheus_memory_limit = 768
# Override health check timeouts, intervals and counts
; [deploy]
; wait_max_sec = 240
; wait_interval = 3
; wait_green_count = 3