-
Notifications
You must be signed in to change notification settings - Fork 38.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
ci-kubernetes-node-kubelet-serial: broken test run #37590
Comments
Multiple broken tests: Failed: [k8s.io] Kubelet Container Manager [Serial] Validate OOM score adjustments once the node is setup pod infra containers oom-score-adj should be -998 and best effort container's should be 1000 {E2eNode Suite}
Issues about this test specifically: #33232 Failed: [k8s.io] Resource-usage [Serial] [Slow] regular resource usage tracking resource tracking for 10 pods per node {E2eNode Suite}
Issues about this test specifically: #30525 #31835 Failed: [k8s.io] Density [Serial] [Slow] create a batch of pods latency/resource should be within limit when create 10 pods with 0 interval {E2eNode Suite}
Issues about this test specifically: #30523 #32022 #33291 #33547 Failed: [k8s.io] InodeEviction [Slow] [Serial] [Disruptive] when we run containers that should cause Disk Pressure due to Inodes should eventually see Disk Pressure due to Inodes, and then evict all of the correct pods {E2eNode Suite}
Failed: [k8s.io] GarbageCollect [Serial] Garbage Collection Test: One Non-restarting Container Should eventually garbage collect containers when we exceed the number of dead containers per container {E2eNode Suite}
Failed: [k8s.io] Restart [Serial] [Slow] [Disruptive] Docker Daemon Network should recover from ip leak {E2eNode Suite}
Failed: [k8s.io] MemoryEviction [Slow] [Serial] [Disruptive] when there is memory pressure should evict pods in the correct order (besteffort first, then burstable, then guaranteed) {E2eNode Suite}
Issues about this test specifically: #32433 |
Multiple broken tests: Failed: [k8s.io] Density [Serial] [Slow] create a batch of pods latency/resource should be within limit when create 10 pods with 0 interval {E2eNode Suite}
Issues about this test specifically: #30523 #32022 #33291 #33547 Failed: [k8s.io] InodeEviction [Slow] [Serial] [Disruptive] when we run containers that should cause Disk Pressure due to Inodes should eventually see Disk Pressure due to Inodes, and then evict all of the correct pods {E2eNode Suite}
Failed: [k8s.io] GarbageCollect [Serial] Garbage Collection Test: One Non-restarting Container Should eventually garbage collect containers when we exceed the number of dead containers per container {E2eNode Suite}
Failed: [k8s.io] Density [Serial] [Slow] create a sequence of pods latency/resource should be within limit when create 10 pods with 50 background pods {E2eNode Suite}
Issues about this test specifically: #30878 #31743 #31877 #32044 |
Failed: https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-node-kubelet-serial/12/
Run so broken it didn't make JUnit output!
The text was updated successfully, but these errors were encountered: