-
Notifications
You must be signed in to change notification settings - Fork 38.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Automated cherry pick of #100183: Mark volume as uncertain after Unmount* fails #107014
Automated cherry pick of #100183: Mark volume as uncertain after Unmount* fails #107014
Conversation
When UnmountDevice fails, kubelet treat the volume mount as uncertain, because it does not know at which stage UnmountDevice failed. It may be already partially unmonted / destroyed. As result, MountDevice will be performer when a new Pod is started on the node after UnmountDevice faiure.
…ere unmounted podVolumesExist() should consider also uncertain volumes (where kubelet does not know if a volume was fully unmounted) when checking for pod's volumes. Added GetPossiblyMountedVolumesForPod for that. Adding uncertain mounts to GetMountedVolumesForPod would potentially break other callers (e.g. `verifyVolumesMountedFunc`).
To know when a volume has been fully unmounted (incl. uncertain mounts).
Change existing desiredStateOfWorldPopulator.findAndAddNewPods tests to use a common initialization function.
desiredStateOfWorldPopulator.findAndRemoveDeletedPods() should remove volumes from DSW when a pod is deleted on the API server and the volume is uncertain in ASW.
@jsafrane: This cherry pick PR is for a release branch and has not yet been approved by Release Managers. To merge this cherry pick, it must first be approved ( AFTER it has been approved by code owners, please ping the kubernetes/release-managers team in a comment to request a cherry pick review. (For details on the patch release process and schedule, see the Patch Releases page.) Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. |
/triage accepted |
/kind bug |
Note: This is already fixed in 1.22 and newer. I tried to backport it all the way to 1.20, but it's too different and it goes EOL soon. |
/lgtm |
/approve |
@Random-Liu could you help approve? |
/approve |
Ping @kubernetes/release-managers for approval. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
/lgtm
[APPROVALNOTIFIER] This PR is APPROVED This pull-request has been approved by: cpanato, jingxu97, jsafrane, sjenning The full list of commands accepted by this bot can be found here. The pull request process is described here
Needs approval from an approver in each of these files:
Approvers can indicate their approval by writing |
Cherry pick of #100183 on release-1.21.
#100183: Add e2e test for NodeUnstage error cases
For details on the cherry pick process, see the cherry pick requests page.