New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Ephemeral storage requests on initContainers can cause pod to be unschedulable #96083
Comments
@jmcmeek: This issue is currently awaiting triage. If a SIG or subproject determines this is a relevant issue, they will accept it by applying the The Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. |
/sig sig-scheduling |
@jmcmeek: The label(s) In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. |
/sig scheduling |
I can confirm it's a bug. Will work on it. /assign |
Clarifying the recreate steps: I disabled the LocalStorageCapacityIsolation feature guide by adding |
Thanks @jmcmeek for confirming! |
What happened:
In a v1.18 cluster with the LocalStorageCapacityIsolation feature gate disabled setting an ephemeral-storage request or limit causes the pod to be unschedulable:
Warning FailedScheduling 56s (x3 over 119s) default-scheduler 0/3 nodes are available: 3 Insufficient ephemeral-storage.
If the ephemeral-storage request is removed the pod is scheduled.
This behavior exists in v1.19 and a recent v1.20 build.
What you expected to happen:
In previous releases such pods are scheduled. I expect that any ephemeral-storage limit is not enforced.
How to reproduce it (as minimally and precisely as possible):
Disable the LocalStorageCapacityIsolation feature guide by adding
--feature-gates=LocalStorageCapacityIsolation=false
to the kube-scheduler.Create a pod like this:
Anything else we need to know?:
The problem was originally seen after moving from IBM Cloud Red Hat OpenShift v4.4 to v4.5. Pods that worked in 4.4 (Kubernetes v1.17) remained in a Pending state in OpenShift v4.5 (Kubernetes 1.18). We discovered that the IBM Cloud OpenShift service has the LocalStorageCapacityIsolation feature-gate disabled while Red Hat has enabled the feature. We then replicated the behavior with a plain Kubernetes cluster.
Environment:
kubectl version
): v1.18.10, v1.19.3, v1.20.0cat /etc/os-release
):uname -a
):The text was updated successfully, but these errors were encountered: