Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

kubernetes-e2e-gke-staging-parallel: broken test run #30066

Closed
k8s-github-robot opened this issue Aug 4, 2016 · 82 comments
Closed

kubernetes-e2e-gke-staging-parallel: broken test run #30066

k8s-github-robot opened this issue Aug 4, 2016 · 82 comments
Assignees
Labels
area/test-infra kind/flake Categorizes issue or PR as related to a flaky test. priority/critical-urgent Highest priority. Must be actively worked on as someone's top priority right now.

Comments

@k8s-github-robot
Copy link

Failed: https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/kubernetes-e2e-gke-staging-parallel/6627/

Run so broken it didn't make JUnit output!

@k8s-github-robot k8s-github-robot added priority/backlog Higher priority than priority/awaiting-more-evidence. area/test-infra kind/flake Categorizes issue or PR as related to a flaky test. labels Aug 4, 2016
@lavalamp
Copy link
Member

lavalamp commented Aug 4, 2016

Hm, not sure why the bot didn't link to the old issue: #28413

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot k8s-github-robot added priority/important-soon Must be staffed and worked on either currently, or very soon, ideally in time for the next release. and removed priority/backlog Higher priority than priority/awaiting-more-evidence. labels Aug 5, 2016
@k8s-github-robot
Copy link
Author

@k8s-github-robot k8s-github-robot added priority/critical-urgent Highest priority. Must be actively worked on as someone's top priority right now. and removed priority/important-soon Must be staffed and worked on either currently, or very soon, ideally in time for the next release. labels Aug 5, 2016
@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/kubernetes-e2e-gke-staging-parallel/7006/

Multiple broken tests:

Failed: [k8s.io] ServiceAccounts should mount an API token into pods [Conformance] {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/service_accounts.go:240
Expected error:
    <*errors.errorString | 0xc82089c1c0>: {
        s: "gave up waiting for pod 'pod-service-account-84e6ecc6-6063-11e6-ae8b-0242ac110004' to be 'success or failure' after 5m0s",
    }
    gave up waiting for pod 'pod-service-account-84e6ecc6-6063-11e6-ae8b-0242ac110004' to be 'success or failure' after 5m0s
not to have occurred

Failed: [k8s.io] [HPA] Horizontal pod autoscaling (scale resource: CPU) [k8s.io] ReplicationController light Should scale from 2 pods to 1 pod {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/horizontal_pod_autoscaling.go:100
Expected error:
    <*errors.errorString | 0xc821069b70>: {
        s: "Only 1 pods started out of 2",
    }
    Only 1 pods started out of 2
not to have occurred

Issues about this test specifically: #27196 #28998

Failed: [k8s.io] [HPA] Horizontal pod autoscaling (scale resource: CPU) [k8s.io] ReplicationController light Should scale from 1 pod to 2 pods {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/horizontal_pod_autoscaling.go:88
Expected error:
    <*errors.errorString | 0xc820c22be0>: {
        s: "Only 0 pods started out of 1",
    }
    Only 0 pods started out of 1
not to have occurred

Issues about this test specifically: #27443 #27835 #28900

Failed: [k8s.io] Deployment deployment should label adopted RSs and pods {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/deployment.go:82
Expected error:
    <*errors.errorString | 0xc820575d80>: {
        s: "failed to wait for pods running: [timed out waiting for the condition]",
    }
    failed to wait for pods running: [timed out waiting for the condition]
not to have occurred

Issues about this test specifically: #29629

Failed: [k8s.io] Kubectl client [k8s.io] Kubectl describe should check if kubectl describe prints relevant information for rc and pods [Conformance] {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl.go:583
Aug 12 01:10:23.816: Verified 0 of 1 pods , error : timed out waiting for the condition

Issues about this test specifically: #28774

Failed: [k8s.io] EmptyDir volumes volume on default medium should have the correct mode [Conformance] {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/empty_dir.go:93
Expected error:
    <*errors.errorString | 0xc8202bd5d0>: {
        s: "gave up waiting for pod 'pod-84fae168-6063-11e6-8f36-0242ac110004' to be 'success or failure' after 5m0s",
    }
    gave up waiting for pod 'pod-84fae168-6063-11e6-8f36-0242ac110004' to be 'success or failure' after 5m0s
not to have occurred

Failed: [k8s.io] Kubectl client [k8s.io] Kubectl run rc should create an rc from an image [Conformance] {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl.go:895
Aug 12 01:10:23.191: Pods for rc e2e-test-nginx-rc were not ready

Issues about this test specifically: #28507 #29315

Failed: [k8s.io] EmptyDir volumes should support (root,0777,default) [Conformance] {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/empty_dir.go:105
Expected error:
    <*errors.errorString | 0xc820a39590>: {
        s: "gave up waiting for pod 'pod-922bab49-6063-11e6-90c2-0242ac110004' to be 'success or failure' after 5m0s",
    }
    gave up waiting for pod 'pod-922bab49-6063-11e6-90c2-0242ac110004' to be 'success or failure' after 5m0s
not to have occurred

Issues about this test specifically: #26780

@k8s-github-robot
Copy link
Author

https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/kubernetes-e2e-gke-staging-parallel/7147/

Multiple broken tests:

Failed: [k8s.io] Services should create endpoints for unready pods {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/service.go:960
Aug 14 23:48:08.767: expected un-ready endpoint for Service webserver within 5m0s, stdout: 

Issues about this test specifically: #26172

Failed: [k8s.io] DNS should provide DNS for pods for Hostname and Subdomain Annotation {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/dns.go:389
Expected error:
    <*errors.errorString | 0xc8200e40b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
not to have occurred

Issues about this test specifically: #28337

Failed: [k8s.io] Kubectl client [k8s.io] Kubectl cluster-info should check if Kubernetes master services is included in cluster-info [Conformance] {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl.go:483
Aug 14 23:39:09.538: Missing KubeDNS in kubectl cluster-info

Issues about this test specifically: #28420

Failed: [k8s.io] DNS should provide DNS for services [Conformance] {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/dns.go:352
Expected error:
    <*errors.errorString | 0xc8200ec0b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
not to have occurred

Issues about this test specifically: #26168 #27450

Failed: [k8s.io] [HPA] Horizontal pod autoscaling (scale resource: CPU) [k8s.io] ReplicationController light Should scale from 1 pod to 2 pods {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/horizontal_pod_autoscaling.go:88
Aug 14 23:58:44.752: timeout waiting 15m0s for pods size to be 2

Issues about this test specifically: #27443 #27835 #28900

Failed: [k8s.io] Kubectl client [k8s.io] Guestbook application should create and stop a working application [Conformance] {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl.go:265
Aug 14 23:49:24.261: Frontend service did not start serving content in 600 seconds.

Issues about this test specifically: #26175 #26846 #27334 #28293 #29149

Failed: [k8s.io] DNS should provide DNS for the cluster [Conformance] {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/dns.go:304
Expected error:
    <*errors.errorString | 0xc8200e40b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
not to have occurred

Issues about this test specifically: #26194 #26338 #30345

Failed: [k8s.io] Networking should provide Internet connection for containers [Conformance] {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/networking.go:55
Expected error:
    <*errors.errorString | 0xc8207a6920>: {
        s: "pod 'wget-test' terminated with failure: &{ExitCode:1 Signal:0 Reason:Error Message: StartedAt:{Time:2016-08-14 23:38:10 -0700 PDT} FinishedAt:{Time:2016-08-14 23:38:40 -0700 PDT} ContainerID:docker://88785b277abf633360c1279aec8d6d1bd91896e800ebe375ada583f8f149f25e}",
    }
    pod 'wget-test' terminated with failure: &{ExitCode:1 Signal:0 Reason:Error Message: StartedAt:{Time:2016-08-14 23:38:10 -0700 PDT} FinishedAt:{Time:2016-08-14 23:38:40 -0700 PDT} ContainerID:docker://88785b277abf633360c1279aec8d6d1bd91896e800ebe375ada583f8f149f25e}
not to have occurred

Issues about this test specifically: #26171 #28188

@k8s-github-robot
Copy link
Author

https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/kubernetes-e2e-gke-staging-parallel/7153/

Multiple broken tests:

Failed: [k8s.io] V1Job should scale a job up {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:133
Aug 15 02:43:04.124: Couldn't delete ns "e2e-tests-v1job-rc21n": namespace e2e-tests-v1job-rc21n was not deleted within limit: timed out waiting for the condition, pods remaining: []

Issues about this test specifically: #29976 #30464

Failed: [k8s.io] Job should scale a job up {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:133
Aug 15 02:43:44.622: Couldn't delete ns "e2e-tests-job-ubpur": namespace e2e-tests-job-ubpur was not deleted within limit: timed out waiting for the condition, pods remaining: []

Issues about this test specifically: #29511 #29987 #30238

Failed: [k8s.io] Services should be able to create a functioning NodePort service {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/service.go:403
Aug 15 02:40:47.410: Failed waiting for pods to be running: Timeout waiting for 1 pods to be ready

Issues about this test specifically: #28064 #28569

Failed: [k8s.io] Pods should be restarted with a /healthz http liveness probe [Conformance] {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/pods.go:1104
Aug 15 02:40:29.304: pod e2e-tests-pods-kzwd5/liveness-http - expected number of restarts: %!t(int=1), found restarts: %!t(int32=0)

@k8s-github-robot
Copy link
Author

1 similar comment
@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/kubernetes-e2e-gke-staging-parallel/7259/

Multiple broken tests:

Failed: [k8s.io] V1Job should fail a job {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:133
Aug 17 07:43:47.546: Couldn't delete ns "e2e-tests-v1job-f195z": the server does not allow access to the requested resource (delete namespaces e2e-tests-v1job-f195z)

Issues about this test specifically: #27704 #30127 #30602

Failed: [k8s.io] ResourceQuota should create a ResourceQuota and capture the life of a secret. {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:132
Expected error:
    <*errors.StatusError | 0xc8202fe680>: {
        ErrStatus: {
            TypeMeta: {Kind: "", APIVersion: ""},
            ListMeta: {SelfLink: "", ResourceVersion: ""},
            Status: "Failure",
            Message: "the server does not allow access to the requested resource (get serviceAccounts)",
            Reason: "Forbidden",
            Details: {
                Name: "",
                Group: "",
                Kind: "serviceAccounts",
                Causes: [
                    {
                        Type: "UnexpectedServerResponse",
                        Message: "Forbidden: \"/api/v1/watch/namespaces/e2e-tests-resourcequota-d2vtt/serviceaccounts?fieldSelector=metadata.name%3Ddefault\"",
                        Field: "",
                    },
                ],
                RetryAfterSeconds: 0,
            },
            Code: 403,
        },
    }
    the server does not allow access to the requested resource (get serviceAccounts)
not to have occurred

Failed: [k8s.io] Downward API volume should provide podname only [Conformance] {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:132
Expected error:
    <*errors.StatusError | 0xc820bd1800>: {
        ErrStatus: {
            TypeMeta: {Kind: "", APIVersion: ""},
            ListMeta: {SelfLink: "", ResourceVersion: ""},
            Status: "Failure",
            Message: "the server does not allow access to the requested resource (get serviceAccounts)",
            Reason: "Forbidden",
            Details: {
                Name: "",
                Group: "",
                Kind: "serviceAccounts",
                Causes: [
                    {
                        Type: "UnexpectedServerResponse",
                        Message: "Forbidden: \"/api/v1/watch/namespaces/e2e-tests-downward-api-0m60i/serviceaccounts?fieldSelector=metadata.name%3Ddefault\"",
                        Field: "",
                    },
                ],
                RetryAfterSeconds: 0,
            },
            Code: 403,
        },
    }
    the server does not allow access to the requested resource (get serviceAccounts)
not to have occurred

Failed: [k8s.io] Pods should support retrieving logs from the container over websockets {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:132
Expected error:
    <*errors.StatusError | 0xc820c5a000>: {
        ErrStatus: {
            TypeMeta: {Kind: "", APIVersion: ""},
            ListMeta: {SelfLink: "", ResourceVersion: ""},
            Status: "Failure",
            Message: "the server does not allow access to the requested resource (get serviceAccounts)",
            Reason: "Forbidden",
            Details: {
                Name: "",
                Group: "",
                Kind: "serviceAccounts",
                Causes: [
                    {
                        Type: "UnexpectedServerResponse",
                        Message: "Forbidden: \"/api/v1/watch/namespaces/e2e-tests-pods-o6uao/serviceaccounts?fieldSelector=metadata.name%3Ddefault\"",
                        Field: "",
                    },
                ],
                RetryAfterSeconds: 0,
            },
            Code: 403,
        },
    }
    the server does not allow access to the requested resource (get serviceAccounts)
not to have occurred

Issues about this test specifically: #30263

Failed: [k8s.io] Job should fail a job {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:133
Aug 17 07:43:46.671: Couldn't delete ns "e2e-tests-job-w2nzh": the server does not allow access to the requested resource (delete namespaces e2e-tests-job-w2nzh)

Issues about this test specifically: #28773 #29506 #30699

Failed: [k8s.io] Deployment paused deployment should be ignored by the controller {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:132
Expected error:
    <*errors.StatusError | 0xc820437a80>: {
        ErrStatus: {
            TypeMeta: {Kind: "", APIVersion: ""},
            ListMeta: {SelfLink: "", ResourceVersion: ""},
            Status: "Failure",
            Message: "the server does not allow access to the requested resource (get serviceAccounts)",
            Reason: "Forbidden",
            Details: {
                Name: "",
                Group: "",
                Kind: "serviceAccounts",
                Causes: [
                    {
                        Type: "UnexpectedServerResponse",
                        Message: "Forbidden: \"/api/v1/watch/namespaces/e2e-tests-deployment-aj6h9/serviceaccounts?fieldSelector=metadata.name%3Ddefault\"",
                        Field: "",
                    },
                ],
                RetryAfterSeconds: 0,
            },
            Code: 403,
        },
    }
    the server does not allow access to the requested resource (get serviceAccounts)
not to have occurred

Issues about this test specifically: #28067 #28378

Failed: [k8s.io] Job should delete a job {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:132
Expected error:
    <*errors.StatusError | 0xc820a0fb00>: {
        ErrStatus: {
            TypeMeta: {Kind: "", APIVersion: ""},
            ListMeta: {SelfLink: "", ResourceVersion: ""},
            Status: "Failure",
            Message: "the server does not allow access to the requested resource (get serviceAccounts)",
            Reason: "Forbidden",
            Details: {
                Name: "",
                Group: "",
                Kind: "serviceAccounts",
                Causes: [
                    {
                        Type: "UnexpectedServerResponse",
                        Message: "Forbidden: \"/api/v1/watch/namespaces/e2e-tests-job-xqa7y/serviceaccounts?fieldSelector=metadata.name%3Ddefault\"",
                        Field: "",
                    },
                ],
                RetryAfterSeconds: 0,
            },
            Code: 403,
        },
    }
    the server does not allow access to the requested resource (get serviceAccounts)
not to have occurred

Issues about this test specifically: #28003

Failed: [k8s.io] LimitRange should create a LimitRange with defaults and ensure pod has those defaults applied. {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:132
Expected error:
    <*errors.StatusError | 0xc820d58900>: {
        ErrStatus: {
            TypeMeta: {Kind: "", APIVersion: ""},
            ListMeta: {SelfLink: "", ResourceVersion: ""},
            Status: "Failure",
            Message: "the server does not allow access to the requested resource (get serviceAccounts)",
            Reason: "Forbidden",
            Details: {
                Name: "",
                Group: "",
                Kind: "serviceAccounts",
                Causes: [
                    {
                        Type: "UnexpectedServerResponse",
                        Message: "Forbidden: \"/api/v1/watch/namespaces/e2e-tests-limitrange-487wl/serviceaccounts?fieldSelector=metadata.name%3Ddefault\"",
                        Field: "",
                    },
                ],
                RetryAfterSeconds: 0,
            },
            Code: 403,
        },
    }
    the server does not allow access to the requested resource (get serviceAccounts)
not to have occurred

Issues about this test specifically: #27503

Failed: [k8s.io] EmptyDir volumes should support (non-root,0666,default) [Conformance] {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:132
Expected error:
    <*errors.StatusError | 0xc82024bf00>: {
        ErrStatus: {
            TypeMeta: {Kind: "", APIVersion: ""},
            ListMeta: {SelfLink: "", ResourceVersion: ""},
            Status: "Failure",
            Message: "the server does not allow access to the requested resource (get serviceAccounts)",
            Reason: "Forbidden",
            Details: {
                Name: "",
                Group: "",
                Kind: "serviceAccounts",
                Causes: [
                    {
                        Type: "UnexpectedServerResponse",
                        Message: "Forbidden: \"/api/v1/watch/namespaces/e2e-tests-emptydir-c2gf6/serviceaccounts?fieldSelector=metadata.name%3Ddefault\"",
                        Field: "",
                    },
                ],
                RetryAfterSeconds: 0,
            },
            Code: 403,
        },
    }
    the server does not allow access to the requested resource (get serviceAccounts)
not to have occurred

Failed: [k8s.io] EmptyDir volumes should support (non-root,0777,tmpfs) [Conformance] {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:132
Expected error:
    <*errors.StatusError | 0xc820c5bc00>: {
        ErrStatus: {
            TypeMeta: {Kind: "", APIVersion: ""},
            ListMeta: {SelfLink: "", ResourceVersion: ""},
            Status: "Failure",
            Message: "the server does not allow access to the requested resource (get serviceAccounts)",
            Reason: "Forbidden",
            Details: {
                Name: "",
                Group: "",
                Kind: "serviceAccounts",
                Causes: [
                    {
                        Type: "UnexpectedServerResponse",
                        Message: "Forbidden: \"/api/v1/watch/namespaces/e2e-tests-emptydir-dcgnd/serviceaccounts?fieldSelector=metadata.name%3Ddefault\"",
                        Field: "",
                    },
                ],
                RetryAfterSeconds: 0,
            },
            Code: 403,
        },
    }
    the server does not allow access to the requested resource (get serviceAccounts)
not to have occurred

@k8s-github-robot
Copy link
Author

https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/kubernetes-e2e-gke-staging-parallel/7260/

Multiple broken tests:

Failed: [k8s.io] [HPA] Horizontal pod autoscaling (scale resource: CPU) [k8s.io] ReplicationController light Should scale from 2 pods to 1 pod {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:132
Expected error:
    <*errors.StatusError | 0xc82084fe80>: {
        ErrStatus: {
            TypeMeta: {Kind: "", APIVersion: ""},
            ListMeta: {SelfLink: "", ResourceVersion: ""},
            Status: "Failure",
            Message: "the server does not allow access to the requested resource (get serviceAccounts)",
            Reason: "Forbidden",
            Details: {
                Name: "",
                Group: "",
                Kind: "serviceAccounts",
                Causes: [
                    {
                        Type: "UnexpectedServerResponse",
                        Message: "Forbidden: \"/api/v1/watch/namespaces/e2e-tests-horizontal-pod-autoscaling-h6lwi/serviceaccounts?fieldSelector=metadata.name%3Ddefault\"",
                        Field: "",
                    },
                ],
                RetryAfterSeconds: 0,
            },
            Code: 403,
        },
    }
    the server does not allow access to the requested resource (get serviceAccounts)
not to have occurred

Issues about this test specifically: #27196 #28998

Failed: [k8s.io] Mesos applies slave attributes as labels {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:132
Expected error:
    <*errors.StatusError | 0xc8208bdf00>: {
        ErrStatus: {
            TypeMeta: {Kind: "", APIVersion: ""},
            ListMeta: {SelfLink: "", ResourceVersion: ""},
            Status: "Failure",
            Message: "the server does not allow access to the requested resource (get serviceAccounts)",
            Reason: "Forbidden",
            Details: {
                Name: "",
                Group: "",
                Kind: "serviceAccounts",
                Causes: [
                    {
                        Type: "UnexpectedServerResponse",
                        Message: "Forbidden: \"/api/v1/watch/namespaces/e2e-tests-pods-7qabb/serviceaccounts?fieldSelector=metadata.name%3Ddefault\"",
                        Field: "",
                    },
                ],
                RetryAfterSeconds: 0,
            },
            Code: 403,
        },
    }
    the server does not allow access to the requested resource (get serviceAccounts)
not to have occurred

Issues about this test specifically: #28359

Failed: [k8s.io] EmptyDir volumes should support (non-root,0666,tmpfs) [Conformance] {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:132
Expected error:
    <*errors.StatusError | 0xc820a78c00>: {
        ErrStatus: {
            TypeMeta: {Kind: "", APIVersion: ""},
            ListMeta: {SelfLink: "", ResourceVersion: ""},
            Status: "Failure",
            Message: "the server does not allow access to the requested resource (get serviceAccounts)",
            Reason: "Forbidden",
            Details: {
                Name: "",
                Group: "",
                Kind: "serviceAccounts",
                Causes: [
                    {
                        Type: "UnexpectedServerResponse",
                        Message: "Forbidden: \"/api/v1/watch/namespaces/e2e-tests-emptydir-bdpga/serviceaccounts?fieldSelector=metadata.name%3Ddefault\"",
                        Field: "",
                    },
                ],
                RetryAfterSeconds: 0,
            },
            Code: 403,
        },
    }
    the server does not allow access to the requested resource (get serviceAccounts)
not to have occurred

Failed: [k8s.io] V1Job should scale a job up {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:132
Expected error:
    <*errors.StatusError | 0xc820527d00>: {
        ErrStatus: {
            TypeMeta: {Kind: "", APIVersion: ""},
            ListMeta: {SelfLink: "", ResourceVersion: ""},
            Status: "Failure",
            Message: "the server does not allow access to the requested resource (get serviceAccounts)",
            Reason: "Forbidden",
            Details: {
                Name: "",
                Group: "",
                Kind: "serviceAccounts",
                Causes: [
                    {
                        Type: "UnexpectedServerResponse",
                        Message: "Forbidden: \"/api/v1/watch/namespaces/e2e-tests-v1job-4m8ey/serviceaccounts?fieldSelector=metadata.name%3Ddefault\"",
                        Field: "",
                    },
                ],
                RetryAfterSeconds: 0,
            },
            Code: 403,
        },
    }
    the server does not allow access to the requested resource (get serviceAccounts)
not to have occurred

Issues about this test specifically: #29976 #30464 #30687

Failed: [k8s.io] MetricsGrabber should grab all metrics from a ControllerManager. {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:132
Expected error:
    <*errors.StatusError | 0xc82055c900>: {
        ErrStatus: {
            TypeMeta: {Kind: "", APIVersion: ""},
            ListMeta: {SelfLink: "", ResourceVersion: ""},
            Status: "Failure",
            Message: "the server does not allow access to the requested resource (get serviceAccounts)",
            Reason: "Forbidden",
            Details: {
                Name: "",
                Group: "",
                Kind: "serviceAccounts",
                Causes: [
                    {
                        Type: "UnexpectedServerResponse",
                        Message: "Forbidden: \"/api/v1/watch/namespaces/e2e-tests-metrics-grabber-64vfj/serviceaccounts?fieldSelector=metadata.name%3Ddefault\"",
                        Field: "",
                    },
                ],
                RetryAfterSeconds: 0,
            },
            Code: 403,
        },
    }
    the server does not allow access to the requested resource (get serviceAccounts)
not to have occurred

Failed: [k8s.io] Kubectl client [k8s.io] Simple pod should support port-forward {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:132
Expected error:
    <*errors.StatusError | 0xc820acdb80>: {
        ErrStatus: {
            TypeMeta: {Kind: "", APIVersion: ""},
            ListMeta: {SelfLink: "", ResourceVersion: ""},
            Status: "Failure",
            Message: "the server does not allow access to the requested resource (get serviceAccounts)",
            Reason: "Forbidden",
            Details: {
                Name: "",
                Group: "",
                Kind: "serviceAccounts",
                Causes: [
                    {
                        Type: "UnexpectedServerResponse",
                        Message: "Forbidden: \"/api/v1/watch/namespaces/e2e-tests-kubectl-mpe7x/serviceaccounts?fieldSelector=metadata.name%3Ddefault\"",
                        Field: "",
                    },
                ],
                RetryAfterSeconds: 0,
            },
            Code: 403,
        },
    }
    the server does not allow access to the requested resource (get serviceAccounts)
not to have occurred

Issues about this test specifically: #28371 #29604

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/kubernetes-e2e-gke-staging-parallel/7435/

Multiple broken tests:

Failed: [k8s.io] Dynamic provisioning [k8s.io] DynamicProvisioner should create and delete persistent volumes {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/volume_provisioning.go:123
Expected error:
    <*errors.errorString | 0xc8207c6c90>: {
        s: "gave up waiting for pod 'pvc-volume-tester-tr7wh' to be 'success or failure' after 15m0s",
    }
    gave up waiting for pod 'pvc-volume-tester-tr7wh' to be 'success or failure' after 15m0s
not to have occurred

Issues about this test specifically: #26682 #28884

Failed: [k8s.io] Downward API volume should provide container's cpu request {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/downwardapi_volume.go:161
Expected error:
    <*errors.errorString | 0xc8208de890>: {
        s: "gave up waiting for pod 'downwardapi-volume-bdbf620d-6769-11e6-b1fe-0242ac11000e' to be 'success or failure' after 5m0s",
    }
    gave up waiting for pod 'downwardapi-volume-bdbf620d-6769-11e6-b1fe-0242ac11000e' to be 'success or failure' after 5m0s
not to have occurred

Failed: [k8s.io] Downward API should provide pod name and namespace as env vars [Conformance] {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/downward_api.go:62
Expected error:
    <*errors.errorString | 0xc820b0be40>: {
        s: "gave up waiting for pod 'downward-api-ce198216-6769-11e6-a18e-0242ac11000e' to be 'success or failure' after 5m0s",
    }
    gave up waiting for pod 'downward-api-ce198216-6769-11e6-a18e-0242ac11000e' to be 'success or failure' after 5m0s
not to have occurred

Failed: [k8s.io] Job should run a job to completion when tasks succeed {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/job.go:59
Expected error:
    <*errors.errorString | 0xc8200e20b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
not to have occurred

Failed: [k8s.io] V1Job should fail a job {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:133
Aug 20 23:48:40.049: Couldn't delete ns "e2e-tests-v1job-mb7wz": namespace e2e-tests-v1job-mb7wz was not deleted within limit: timed out waiting for the condition, pods remaining: [foo-irwj6]

Issues about this test specifically: #27704 #30127 #30602

Failed: [k8s.io] Job should scale a job down {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/job.go:162
Expected error:
    <*errors.errorString | 0xc8200e20b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
not to have occurred

Issues about this test specifically: #29066 #30592 #31065

Failed: [k8s.io] Kubectl client [k8s.io] Simple pod should support exec {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl.go:311
Expected error:
    <*errors.errorString | 0xc820b19490>: {
        s: "Error running &{/workspace/kubernetes/platforms/linux/amd64/kubectl [kubectl --server=https://104.198.199.212 --kubeconfig=/workspace/.kube/config exec --namespace=e2e-tests-kubectl-m712n -i nginx bash] []  0xc8203f4c00  Error from server: ssh: rejected: administratively prohibited (open failed)\n [] <nil> 0xc8207cc160 exit status 1 <nil> true [0xc8203f4c00 0xc8203f4c48 0xc8203f4cb8] [0xc8203f4c48 0xc8203f4cb8] [0xc8203f4c40 0xc8203f4ca8] [0xa96280 0xa96280] 0xc820b85800}:\nCommand stdout:\n\nstderr:\nError from server: ssh: rejected: administratively prohibited (open failed)\n\nerror:\nexit status 1\n",
    }
    Error running &{/workspace/kubernetes/platforms/linux/amd64/kubectl [kubectl --server=https://104.198.199.212 --kubeconfig=/workspace/.kube/config exec --namespace=e2e-tests-kubectl-m712n -i nginx bash] []  0xc8203f4c00  Error from server: ssh: rejected: administratively prohibited (open failed)
     [] <nil> 0xc8207cc160 exit status 1 <nil> true [0xc8203f4c00 0xc8203f4c48 0xc8203f4cb8] [0xc8203f4c48 0xc8203f4cb8] [0xc8203f4c40 0xc8203f4ca8] [0xa96280 0xa96280] 0xc820b85800}:
    Command stdout:

    stderr:
    Error from server: ssh: rejected: administratively prohibited (open failed)

    error:
    exit status 1

not to have occurred

Issues about this test specifically: #28426

Failed: [k8s.io] Job should scale a job up {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/job.go:137
Expected error:
    <*errors.errorString | 0xc82007bf80>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
not to have occurred

Issues about this test specifically: #29511 #29987 #30238

Failed: [k8s.io] V1Job should scale a job up {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/batch_v1_jobs.go:141
Expected error:
    <*errors.errorString | 0xc8200db060>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
not to have occurred

Issues about this test specifically: #29976 #30464 #30687

Failed: [k8s.io] Proxy version v1 should proxy through a service and a pod [Conformance] {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/proxy.go:261
Expected error:
    <*errors.errorString | 0xc820bfe370>: {
        s: "Only 0 pods started out of 1",
    }
    Only 0 pods started out of 1
not to have occurred

Issues about this test specifically: #26164 #26210

Failed: [k8s.io] Generated release_1_2 clientset should create pods, delete pods, watch pods {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/generated_clientset.go:170
Expected error:
    <*errors.errorString | 0xc8200e20b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
not to have occurred

Failed: [k8s.io] [HPA] Horizontal pod autoscaling (scale resource: CPU) [k8s.io] ReplicationController light Should scale from 1 pod to 2 pods {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/horizontal_pod_autoscaling.go:88
Aug 20 23:51:42.985: timeout waiting 15m0s for pods size to be 2

Issues about this test specifically: #27443 #27835 #28900

Failed: [k8s.io] Job should keep restarting failed pods {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:133
Aug 20 23:58:01.908: Couldn't delete ns "e2e-tests-job-tkztw": namespace e2e-tests-job-tkztw was not deleted within limit: timed out waiting for the condition, pods remaining: [all-fail-06mud]

Issues about this test specifically: #28006 #28866 #29613

Failed: [k8s.io] kubelet [k8s.io] Clean up pods on node kubelet should be able to delete 10 pods per node in 1m0s. {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubelet.go:223
Expected error:
    <*errors.errorString | 0xc8209b0a60>: {
        s: "Only 21 pods started out of 30",
    }
    Only 21 pods started out of 30
not to have occurred

Issues about this test specifically: #28106

Failed: [k8s.io] ReplicaSet should serve a basic image on each replica with a private image {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/replica_set.go:47
Expected error:
    <*errors.errorString | 0xc8200d60b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
not to have occurred

Failed: [k8s.io] Kubectl client [k8s.io] Update Demo should scale a replication controller [Conformance] {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl.go:223
Aug 20 23:47:48.474: Timed out after 300 seconds waiting for name=update-demo pods to reach valid state

Issues about this test specifically: #28437 #29084 #29256 #29397

Failed: [k8s.io] V1Job should keep restarting failed pods {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:133
Aug 20 23:49:18.616: Couldn't delete ns "e2e-tests-v1job-mb7wz": Operation cannot be fulfilled on namespaces "e2e-tests-v1job-mb7wz": The system is ensuring all content is removed from this namespace.  Upon completion, this namespace will automatically be purged by the system.

Issues about this test specifically: #29657

Failed: [k8s.io] V1Job should run a job to completion when tasks succeed {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/batch_v1_jobs.go:63
Expected error:
    <*errors.errorString | 0xc8200d80b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
not to have occurred

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@k8s-github-robot
Copy link
Author

@fejta
Copy link
Contributor

fejta commented Aug 22, 2016

I suspect we're having agent cpu issues

https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/kubernetes-e2e-gke-staging-parallel/7498/

++ curl -fsS --retry 3 https://raw.githubusercontent.com/kubernetes/test-infra/master/jenkins/dockerized-e2e-runner.sh
curl: (7) couldn't connect to host

@spxtr
Copy link
Contributor

spxtr commented Aug 22, 2016

Was this due to the GCE outage this morning?

@fejta
Copy link
Contributor

fejta commented Aug 22, 2016

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
area/test-infra kind/flake Categorizes issue or PR as related to a flaky test. priority/critical-urgent Highest priority. Must be actively worked on as someone's top priority right now.
Projects
None yet
Development

No branches or pull requests

4 participants