-
Notifications
You must be signed in to change notification settings - Fork 4.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
running "kubectl drain" no longer removes machine instances from ELBs #10774
Comments
If I understand you correctly, you mean that an ELB provisioned through a Furthermore, kOps have not changed anything that will influence the behavior of commands like |
Thanks for your quick reply. Just to clarify - the behavior I'm seeing is that an ELB provisioned through a Ok, it makes sense that this change isn't related to kops. Considering this behavior how do you recommend updating instances gracefully with kops? As far as I can tell, running |
If the problem is that proxied connections are killed there are two solutions that I am aware of: |
Thanks for your suggestions! I went down a cilium/eBPF/maglev rabbit hole and found it very interesting. Now I think I have a better handle on the issue I described earlier but while testing out NLB's I ran into a bigger issue. In my setup, when the master node's machine instance restarts, the NLB target instances all get deregistered simultaneously and then re-registered which causes a long outage. I realize this isn't a kops issue but would you happen to know if this is expected behavior and if so how to prevent it? |
A master instance restart should not cause the entire NLB target group to re-register. Is that something you see happening consistently? |
Yes, consistently. I tried it about five times using different versions of K8S (1.19, 1.20 and 1.21) and saw it happen each time. I noticed it first when running |
Managing those things is handled by the control plane. I don't think it would actully de-register nodes just because ... but maybe there is a timeout or something that causes this. Do you have the chance to test a HA cluster as well? |
I just tested it on a 3-master HA cluster and found that terminating 1/3 or 2/3 master instances simultaneously did not de-register the nodes but terminating 3/3 simultaneously did. |
Kubernetes bug report for reference kubernetes/kubernetes#100779 |
Issues go stale after 90d of inactivity. If this issue is safe to close now please do so with Send feedback to sig-contributor-experience at kubernetes/community. |
Stale issues rot after 30d of inactivity. If this issue is safe to close now please do so with Send feedback to sig-contributor-experience at kubernetes/community. |
The Kubernetes project currently lacks enough active contributors to adequately respond to all issues and PRs. This bot triages issues and PRs according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /close |
@k8s-triage-robot: Closing this issue. In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. |
1. What
kops
version are you running? The commandkops version
, will displaythis information.
1.19.0
2. What Kubernetes version are you running?
kubectl version
will print theversion if a cluster is running or provide the Kubernetes version specified as
a
kops
flag.1.19.7
3. What cloud provider are you using?
AWS
4. What commands did you run? What is the simplest way to reproduce this issue?
kubectl drain
5. What happened after the commands executed?
Previously, using kops/k8s 1.18.2/1.18.10 running
kubectl drain
would remove machine instances from classic ELBs. Now runningkubectl drain
does not do so.6. What did you expect to happen?
I expected running
kubectl drain
to remove the node machine instances from classic ELBs. I was also hoping that runningkubectl drain
would remove instances from network ELBs but that isn't happening either (with both kops/k8s 1.18.2/1.18.10 and 1.19.0/1.19.7).7. Please provide your cluster manifest. Execute
kops get --name my.example.com -o yaml
to display your cluster manifest.You may want to remove your cluster name and other sensitive information.
8. Please run the commands with most verbose logging by adding the
-v 10
flag.Paste the logs into this report, or in a gist and provide the gist link here.
9. Anything else do we need to know?
The text was updated successfully, but these errors were encountered: