-
Notifications
You must be signed in to change notification settings - Fork 306
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG] configMap volumeMount -- failed to create containerd container: rootpath on mountPath ... The system cannot find the path specified. #3117
Comments
|
I am also facing a similar issue after upgrading to 1.23.5 from 1.22.11. Where windows pods are not getting started and runs into an infinite crash loop with the below errors
|
@andyzhangx I tried re-imaging the vmss instances as well. But it's of no use. |
The I have since created a fresh cluster in a new resource group in a different region and started out with version |
@AbelHu do you have any insight about windows container creation failure on windows node with below error:
|
@andyzhangx it does not contain enough info. Suggest collecting full nodes logs for advanced investigation. You may find the clue in kubelet logs |
@mbelt if you are using AKS managed CSI driver, the identity CSI driver using is For the first and third issues, it's all related to windows container creation failure on Windows node, pls file an azure support ticket. |
@andyzhangx Ack on one and three. For the second issue, no I haven't touched |
Kubelet logs contain one additional piece of information. On nodes where the failing Pod is placed I am seeing:
There are ~5x as many instances of the second error than the first, but the first does not occur on nodes where this pod is not assigned. |
@mbelt I came across this where some changes need to be done on docker file while using containerd containerd/containerd#6300 (comment). Try it if it's of any help. |
@MageshSrinivasulu Confirmed the root cause of the container failing to start was containerd #5067. The other error messages persist, but they must be unrelated. |
Describe the bug
After upgrading from 1.22.6 to 1.23.8
az aks upgrade ... --kubernetes-version 1.23.8
our Windows nodepool made the jump from dockerd to containerd.
Afterwards multiple containers fail to start, csi-drivers on windows nodepools are restarting with errors.
First symptom:
ContainerError while setting up mounted configMap.
On the node, the directory
C:\Windows\TEMP\ctd-volumectd-volume3189089798/620
does exist, a symlink to\\?\Volume{guid}
, but its contents could not be listed.Attempted to mitigate the issue by:
No success
Second Symptom
Errors and restarts of csi-azurefile-node-win-xxxx and csi-azuredisk-node-win-xxxx
Excerpts from csi-azuredisk-node-win:
Full log
The csi-azurefile-node-win pod has the same errors.
The csi-azurexxxx-node pods on the linux VMSS also have the error about failing to parse the resource ID, but none of the other errors.
Third symptom
__An pod from a different deployment is failing to start after the upgrade. __
These errors all relate to storage driver on the node in some way, so I have grouped them in a single bug report.
Environment
The text was updated successfully, but these errors were encountered: