-
Notifications
You must be signed in to change notification settings - Fork 38.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[k8s.io] Load capacity [Feature:Performance] should be able to handle 30 pods per node {Kubernetes e2e suite} #31589
Comments
https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/kubernetes-kubemark-500-gce/5357/ Failed: [k8s.io] Load capacity [Feature:Performance] should be able to handle 30 pods per node {Kubernetes e2e suite}
|
I think there are two different issues here - I thikn that I already debugged the second case.
As a result, we will end up with "too old resource version" error coming from watch (since it was called ~5 seconds later then the get for it). However, this test is pretty big, so the solution for this problem is to actually make the size of the sliding window larger (we can afford it in large clusters). Will send a PR for it. |
@gmarek ^^ |
Automatic merge from submit-queue Increase cache size for RCs Ref #31589 [This should also help with failures of kubemark-scale.]
https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/kubernetes-kubemark-500-gce/5384/ Failed: [k8s.io] Load capacity [Feature:Performance] should be able to handle 30 pods per node {Kubernetes e2e suite}
|
We should double check, but my hypothesis is that it may (similarly to our failures of kubemark-scale) be a consequence of overloaded Jenkins machine. |
I checked the last failure and it's pretty obvious that the machine where the test is running is overloaded. Get() and Update() are the method visible in apiserver logs. And here are the logs (corresponding get and put operations for one of RCs):
As you can see, there are even ~20s breaks between those two consecutive calls. |
https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/kubernetes-kubemark-500-gce/5412/ Failed: [k8s.io] Load capacity [Feature:Performance] should be able to handle 30 pods per node {Kubernetes e2e suite}
|
Hmm - we are now running tests on exclusive machines, and the symptoms are pretty much the same. The new hypothesis is that maybe client (from test) is being throttled? |
Yeah - I confirmed running large kubemark on my own cluster, that the problem is actually throttling in the e2e test client. |
Automatic merge from submit-queue Higher client qps in load Fix #31589
Automatic merge from submit-queue Make it possible to set higher limit for in-flight requests in test Ref #31589
https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/kubernetes-kubemark-500-gce/5336/
Failed: [k8s.io] Load capacity [Feature:Performance] should be able to handle 30 pods per node {Kubernetes e2e suite}
Previous issues for this test: #26544 #26938 #27595 #30146 #30469 #31374 #31427 #31433
The text was updated successfully, but these errors were encountered: