-
Notifications
You must be signed in to change notification settings - Fork 49
Persistent volumes in stuck state after reboot #191
Comments
I'm not sure hat that relates to CLUO? Are you running its Pods with volumes? |
No, not running CLUO with volumes. Is it possible that when the operator terminates the pods, it reboots before PV are properly detached? |
Hey @dustinmm80, are you by chance using the CSI implemention of EBS volumes? I see something similar with the Cinder CSI driver and I suspect it to be related to Just wanted to check in before investigating this. |
Thinking about it, I'm not sure it's only related to CSI. But it's probably a major issue with StatefulSets, because Kubernetes won't create a new And only after scheduling the new StatefulSet pod to a new node the CSI components will start churning and update the |
We had the same issue a few weeks ago - it turned out to be an issue with newer, Nitro instances (c5, t3, ...) - AWS investigated and claims to have now solved the problem. Were you seeing that problem with nitro instances? Are you still seeing the problem? |
It appears there is a race condition when using persistent volumes, where the pod is deleted and the node is rebooted, but the attached volume is still in the process of detaching. Once this happens, the persistent volume is stuck and must be manually removed and recreated.
I'm seeing this on AWS with EBS volumes.
The text was updated successfully, but these errors were encountered: