-
Notifications
You must be signed in to change notification settings - Fork 1.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
calico-kube-controllers and calico-node are not ready (CrashLoopBackOff) #4935
Comments
@HoseinGhanbari I wonder if you might be hitting this issue, which was fixed by increasing the timeout on liveness probes: #4911 Could you share the logs from your kube-controllers and calico/node pods so I can see if there are any obvious errors? |
BTW This is so weird... Let me know if you need some other logs.
|
Fortunately increasing
@caseydavenport Thank you very much |
How can we solve this when we deploy the tigera crd's? I dont see any option. Editing the deployment or daemonset is not a feasible option in a CI/CD pipeline right? |
What is the impact of increasing the timeouts ? |
I deployed a brand new k8s cluster using kubespray, everything works fine but all of the calico related pods are not ready. And after many hours of debugging I couldn't find the reason why calico pods are crashing. I even disabled/stopped the entire firewalld service but nothing changed.
One other important thing is that
calicoctl node status
output is not stable and every time gets called show something different:Another log that shown up often is the following message:
Also tried changing IP_AUTODETECTION_METHOD with each of the following but nothing changed:
Expected Behavior
All pods, daemonset, deployment and replicaset related to calico should be in READY state.
Current Behavior
All pods, daemonset, deployment and replicaset related to calico is in NOT READY state.
Possible Solution
Nothing yet, I am asking for help on how to debug / overcome this issue.
Steps to Reproduce (for bugs)
Its the latest version of kubespray with the following Context & Environment.
Context
I'm trying to deploy a k8s cluster which has one master and one worker node. Also note that the servers taking part in this cluster are located in an almost airgapped/offline enviroment with limited access to global internet, of course the ansible process of deploying cluster using kubespray was successful but I'm facing this issue with calico pods.
Your Environment
The text was updated successfully, but these errors were encountered: