-
Notifications
You must be signed in to change notification settings - Fork 39.3k
-
Notifications
You must be signed in to change notification settings - Fork 39.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Azure Disks occasionally mounted in a way leading to I/O errors #71453
Comments
In my current test environment:
Notice the mismatch between the disk referenced by lun0 (
|
After deleting the failing Pod and waiting for its successful recreation:
This time disks are matching (
|
/assign |
What happened:
Azure Disks get randomly mounted in a way that makes them unusable, returning
Input/output error
for every disk I/O operation. Affected Pods remain inCrashLoopBack
until someone manually recreates them.What you expected to happen:
Azure Disks are consistently usable by Pods inside which they are mounted.
How to reproduce it:
Create the following StatefulSet. Adjust the replicas and delete the associated Pods (
kubectl delete pod -l app=azure-disk-failure
) until the problem occurs (i.e. some Pod ends up inCrashLoopBackOff
).sts.yaml
Default StorageClass (as created by ACS-Engine):
sc.yaml
Anything else we need to know?:
Nodes were updated from 1.10.4 to 1.10.10, then from 1.10.10 to 0180b22 without a reboot.
According to this document written by @andyzhangx the issue is fixed, but I can still observe it quite often.
Ref Azure/acs-engine#1918
Environment:
/kind bug
/sig storage
/sig azure
The text was updated successfully, but these errors were encountered: