-
Notifications
You must be signed in to change notification settings - Fork 7.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Managed Istio: Manager exited non-zero #27509
Comments
@ostromart this looks like a crash failure. Can you take a look? |
@luiof just a brief look. It appears as if leader election, which is used to ensure only one operator is running at a time within a namespace (aka fencing), is lost because of a timeout reading from the K8s API server. Looking at the K8s client code, the default is 10 seconds for 10 seconds may be insufficient on a heavily loaded K8s API server. @ostromart how do you think we should proceed? Increase the timeout to something more reasonable like 30 seconds? @morvencao if you're bored :) Cheers, |
can add a new env to the template to customize renewDeadline and pass in the manager config |
Any update or eta about the new customization to mitigate the timeout?
|
@richardwxn I made this a P0 for 1.8 as this seems to be an important bug fix. Looks like you have an idea about how to fix it so temporarily assign to you. Please downgrade if you think this is not a P0 or reassign if necessary. thanks! |
This let us set the RENEW_DEADLINE but the leaseDuration limits the renewDeadline. We set the RENEW_DEADLINE to 30s and get this error when the operator pod tries to start.
It seems like the leaseDuration defaults to 15s cc @richardwxn |
Hi team! We have the managed istio on multiple IKS clusters.
We are using the latest 1.7.2 version and sporadically on some clusters we can see logs as the following:
and as:
This seems to appear only some days and everything seems to be fine (the system works as usually).
The
istioctl analyze -n namespace
doesn't show any problem.The
istioctl version
shows that every proxy is updated to latest 1.7.2 version.Any ideas? Are there action that we can take to avoid those crash?
If required, I can attach some extra logs.
The text was updated successfully, but these errors were encountered: