New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
kubelet consistently timing out on attempting to Destroy cgroups #92766
Labels
kind/bug
Categorizes issue or PR as related to a bug.
priority/important-soon
Must be staffed and worked on either currently, or very soon, ideally in time for the next release.
sig/node
Categorizes an issue or PR as relevant to SIG Node.
Milestone
Comments
k8s-ci-robot
added
the
needs-sig
Indicates an issue or PR lacks a `sig/foo` label and requires one.
label
Jul 2, 2020
/sig node |
k8s-ci-robot
added
sig/node
Categorizes an issue or PR as relevant to SIG Node.
and removed
needs-sig
Indicates an issue or PR lacks a `sig/foo` label and requires one.
labels
Jul 2, 2020
I have tried to set |
related #92521 |
it is a regression in runc: opencontainers/runc#2503 |
liggitt
added
the
priority/important-soon
Must be staffed and worked on either currently, or very soon, ideally in time for the next release.
label
Jul 5, 2020
PR opened here: #92862 CRI-O tests: cri-o/cri-o#3858 |
giuseppe
added a commit
to giuseppe/kubernetes
that referenced
this issue
Jul 9, 2020
when the systemd cgroup manager is used, controllers not handled by systemd are created manually afterwards. libcontainer didn't correctly cleanup these cgroups that were leaked on cgroup v1. Closes: kubernetes#92766 Signed-off-by: Giuseppe Scrivano <gscrivan@redhat.com>
sharadg
pushed a commit
to sharadg/kubernetes
that referenced
this issue
Oct 23, 2020
when the systemd cgroup manager is used, controllers not handled by systemd are created manually afterwards. libcontainer didn't correctly cleanup these cgroups that were leaked on cgroup v1. Closes: kubernetes#92766 Signed-off-by: Giuseppe Scrivano <gscrivan@redhat.com>
6 tasks
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Labels
kind/bug
Categorizes issue or PR as related to a bug.
priority/important-soon
Must be staffed and worked on either currently, or very soon, ideally in time for the next release.
sig/node
Categorizes an issue or PR as relevant to SIG Node.
What happened:
the CRI-O CI tests have been in a bad shape recently. In debugging, I have found that the kubelet logs are filled with:
AFAICT, this is from a combination of bumping to go 1.14.4, and a03db63
as I ran two different PRs that dropped each of these commits, and there were no similar problems.
I am fairly certain this is NOT a problem with kubernetes directly, but rather some odd interaction between go 1.14 and either libcontainer, go-systemd, or godbus. but I figure it can be opened here to start the conversation
What you expected to happen:
StopUnit should not time out
How to reproduce it (as minimally and precisely as possible):
run a node with cgroupv1
build hyperkube with go 1.14.4 (as is now required)
run
hack/local-up.sh
create and remove a pod and see the cgroup be failed to be torn down
Anything else we need to know?:
Environment:
kubectl version
):master
aws
cat /etc/os-release
):though this also happens on our RHEL 7 boxes
uname -a
):build locally
The text was updated successfully, but these errors were encountered: