Closed
Description
Hello,
We are running a 5 node etcd cluster:
$ etcdctl endpoint status
etcd01, 507905ef22a349ce, 3.4.7, 3.1 GB, true, false, 785, 17694912855, 17694912855,
etcd02, 96622104eaa8652d, 3.4.7, 3.1 GB, false, false, 785, 17694912881, 17694912880,
ectd03, e91fce12ee84c080, 3.4.7, 3.1 GB, false, false, 785, 17694912903, 17694912903,
etcd04, 400fc14411f50272, 3.4.7, 3.1 GB, false, false, 785, 17694912989, 17694912985,
etcd05, 87c46f0b178dc777, 3.4.7, 3.1 GB, false, false, 785, 17694913043, 17694913028,
And we're having some weird performance issue eg:
# etcdctl endpoint health
etcd01 is healthy: successfully committed proposal: took = 12.462058ms
etcd03 is healthy: successfully committed proposal: took = 18.826686ms
etcd02 is healthy: successfully committed proposal: took = 19.418745ms
etcd04 is healthy: successfully committed proposal: took = 24.314474ms
etcd05 is healthy: successfully committed proposal: took = 244.761598ms
# etcdctl endpoint health
etcd01 is healthy: successfully committed proposal: took = 13.505405ms
etcd03 is healthy: successfully committed proposal: took = 21.905048ms
etcd04 is healthy: successfully committed proposal: took = 22.569332ms
etcd02 is healthy: successfully committed proposal: took = 23.10597ms
etcd05 is healthy: successfully committed proposal: took = 24.182998ms
# etcdctl endpoint health
etcd05is healthy: successfully committed proposal: took = 24.854541ms
etcd01 is healthy: successfully committed proposal: took = 86.045049ms
etcd03 is healthy: successfully committed proposal: took = 171.771975ms
etcd04 is healthy: successfully committed proposal: took = 576.218846ms
etcd02 is healthy: successfully committed proposal: took = 1.06666032s
Not sure how to debug it, it looks pretty random. Feel free to ask for more info!