-
Notifications
You must be signed in to change notification settings - Fork 2.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Rancher logs spammed with "Updating workload [ingress-nginx/nginx-ingress-controller]" and "Updating service [frontend] with public endpoints" on a rollback #35798
Comments
Might be tangentially related to: https://github.com/rancher/rancher/issues/35690 |
Seeing this on HA as well. Upgraded from 2.6.2 --> 2.6-head
|
This is seen on a docker install upgrade + rollback 2.6.2 to 2.6-head to 2.6.2 |
I would expect to see this issue on rollback but after the cluster agent has been rolled back to the 2.6.2 agent the messages should stop. Is that the case? If 2.6.2 is running the cluster agent from 2.6.3 there would be an issue, but 2.6.2 is supposed to update the cluster agent back to 2.6.2 eventually. |
@ibuildthecloud Yes, the logs do stop after a few minutes. And would be explained by your comments |
I hit this same issue, but on an upgrade from v2.6.2 to v2.6.3, so maybe we should consider fixing it. It did eventually stop, but went on for 80 minutes and generated 17MB logs with 24,468 entries. Could be related to rancher managing downstream clusters with rancher. |
same issue with v2.6.3 |
@sowmyav27 Why is this closed? After upgrade to 2.6.3 rancher log is full with Updating workload. 2022/01/18 20:48:34 [INFO] Updating workload [cattle-monitoring-system/rancher-monitoring-prometheus-node-exporter] with public endpoints [[{"nodeName":.................... |
same question here, after upgraded from 2.6.2 to 2.6.3, the rancher pod keeps updating the nginx-ingress-controller public endpoints like crazy, and eventually caused panic and restarted.
|
@deniseschannon please reopen this issue, this is still a problem and also etcd size is growing very fast after upgrade to 2.6.3. This is because of Updating workload issue. We can't upgrade to 2.6.3 because of this error and etcd size is growing like crazy. This problem is on all daemonsets, like node-exporter.. |
If you're seeing this issue with workloads and either service or ingress, you should see if you service or ingress is being repeatedly updated with different values. For example this can happen when ingress controllers assign new IP's to an ingress object, and another one does the same, so they waffle back and forth. Rancher will take each update and update various objects with that data. This can lead to degraded kube api performance and lots of log messages. |
Rancher Server Setup
e4db3e
to 2.5.11Information about the Cluster
Describe the bug
The text was updated successfully, but these errors were encountered: