Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Iptables sporadically errs out on master #23967

Closed
bprashanth opened this issue Apr 7, 2016 · 1 comment
Closed

Iptables sporadically errs out on master #23967

bprashanth opened this issue Apr 7, 2016 · 1 comment
Labels
sig/network Categorizes an issue or PR as relevant to SIG Network. sig/node Categorizes an issue or PR as relevant to SIG Node.

Comments

@bprashanth
Copy link
Contributor

It is weird. This failure:

https://pantheon.corp.google.com/storage/browser/kubernetes-jenkins/pr-logs/pull/22989/kubernetes-pull-build-test-e2e-gce/32669/artifacts/e2e-gce-builder-2-0-master/
#22989 (comment)

happened because the apiserver didn't come up in 300s, which happened because:

NodeHasSufficientDisk
I0315 14:32:55.787839    3394 kubelet.go:1137] Unable to register e2e-gce-builder-2-0-master with the apiserver: Post https://e2e-gce-builder-2-0-master/api/v1/nodes: dial tcp 10.240.0.2:443: connection refused
E0315 14:32:56.029664    3394 generic.go:195] GenericPLEG: Unable to retrieve pods: cannot connect to Docker endpoint
E0315 14:32:57.030062    3394 generic.go:195] GenericPLEG: Unable to retrieve pods: cannot connect to Docker endpoint
I0315 14:32:57.843326    3394 kubelet.go:2365] skipping pod synchronization - [Error on adding ip table rules: exit status 2 container runtime is down]

So essentially the Kubelet never setup cbr0 but did restart docker, which failed to startup because cbr0 doesn't exist.

@bprashanth
Copy link
Contributor Author

So essentially the Kubelet never setup cbr0 but did restart docker, which failed to startup because cbr0 doesn't exist.

Actually how did this happen? it must've been a spurious docker restart (because we only pkill docker after configuring iptables rules, which is failing).

Also I noticed it was failing in the same way on all nodes, so not sure what's going on at this point. Maybe it was a real kubelet bug that we fixed (but iptables exiting with 2 is indicative of, eg, wrong formatting of rules and we haven't touched the rule in question in a while, right?

func ensureIPTablesMasqRule(nonMasqueradeCIDR string) error {
)

@thockin thockin added the sig/network Categorizes an issue or PR as relevant to SIG Network. label May 19, 2017
@thockin thockin closed this as completed May 19, 2017
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
sig/network Categorizes an issue or PR as relevant to SIG Network. sig/node Categorizes an issue or PR as relevant to SIG Node.
Projects
None yet
Development

No branches or pull requests

2 participants