-
Notifications
You must be signed in to change notification settings - Fork 38.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
image-gc-high-threshold should be lower than value causing hard eviction nodefs.available or imagefs.available #124868
Comments
This issue is currently awaiting triage. If a SIG or subproject determines this is a relevant issue, they will accept it by applying the The Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes-sigs/prow repository. |
/sig node |
"imageGCHighThresholdPercent" used by realImageGCManager that just delete no used images. |
@olderTaoist maybe i got the idea you thought ... you mean, that without running eviction, there will be no unused images to garbage collect. but there could be lots of images not used anymore, because of cronjobs, previous image versions, etc. in reality with this setting i eliminated lots of situations where node is marked with disk pressure for short time. and just moving workload betwen nodes to garbage collect unused here, but redownload on other node is not ideal. better to proactively clean up some space and keep workload going. btw in the worst scenario, if it does not prune enough, eviction still happens and it will work the way it works now. |
i got. It is necessary to modify the default value of |
/assign |
/priority backlog |
What happened?
when disk fills, it hits the hard eviction threshold, causing node disk pressure in the same moment imagegc spots it should prune something and start acting. this casues node going into disk pressure and evicting pods and not just start imagegc soon enough
What did you expect to happen?
i expect imagegc spots the filling disk soon enough, to start garbage collection and empty disk before node hits disk pressure.
from current documentation:
https://kubernetes.io/docs/reference/command-line-tools-reference/kubelet/
and
https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/
https://kubernetes.io/docs/concepts/scheduling-eviction/node-pressure-eviction/#hard-eviction-thresholds
vs.
... this does not make sense having the same percentage for imagefs.available (100-15 = 85 :-)
better approach would be having default values a little bit shifted with eqivalent to setting
... after setting this, i almost never get node disk pressure, because garbage collection and pruning disk happens soon enough
How can we reproduce it (as minimally and precisely as possible)?
fill disk, spot node disk pressure state at the same moment it starts garbage collecting.
Anything else we need to know?
No response
Kubernetes version
this is not version specific as checked documentation at the moment, its for years the same.
Cloud provider
none - kubeadm installation
OS version
not relevant
Install tools
Container runtime (CRI) and version (if applicable)
Related plugins (CNI, CSI, ...) and versions (if applicable)
The text was updated successfully, but these errors were encountered: