New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Node in NotReady status after kernel upgrade and reboot on Ubuntu Vivid (15.04) due to Docker service failure (no aufs) #14162
Comments
Thanks for reporting the issue to kick out more discussion. You just opened a can of worms here :-)
You can upgrade your docker version through salt, or update docker flag to use devicemapper through salt.
We should do better job on defining support metrics, specifying minimal system requirements, running integration tests / end-to-end tests through something like version manager which controls each component's versions, etc. I haven't listed all we should do yet, but you can see there are bunch of pending works here. This is open source project, and I believe our community can help with this.
When we cut Kubernetes v1.0 release, Docker is on 1.7.1, but has several serious issues unresolved, especially on host network. That is why we decided to cut 1.0 with docker v 1.6.2. But there is always a way to upgrade to a new version through salt for cluster admin. For 1.1 release, the plan is cutting it with Docker 1.8.2 . |
It looks like the fix is to change |
Just ran into this issue, submitting a PR to address |
+1, running Kubernetes v1.0.6 production release on AWS, provisioning Vivid instances. |
@four43 fixed in the PR, or install linux-image-virtual on all nodes |
Yup, I commented over there too :) Thanks @paralin |
Fixes AWS ubuntu deployment due to extra-$(uname) vs extra-virtual package being installed. See issue kubernetes#14162 Signed-off-by: Christian Stewart <christian@paral.in>
Hi. I'm writing this partially because I want to start a discussion and to leave some trails for others potentially affected.
I started my journey with Kubernetes just a few days ago. I downloaded the 1.0.6 release (now I know this a only a pre-release and I should've downloaded the 1.0.4 release but this doesn't seem to be related to my particular issue).
I brought up a tiny k8s cluster with the following config:
My cluster was created properly and after a bit of playing I left it in a good shape.
The next day I wanted to investigate some DNS related issue on k8s nodes so I logged in on one of the minions. I noticed
*** System restart required ***
in the MOTD. I ignored it for a moment and investigated the DNS matters (somenslookup
s, nothing dangerous). Then I rebooted the machine.Once the machine was up again I found out that
kubectl
reported it asNotReady
. I started digging and investigating. It took me some time to find out that the Docker service on the node was not running:Further investigation led me to a conclusion that the problem arose from the fact that before the reboot the kernel was upgraded to 3.19.0-28, which no longer by default supports AUFS module which Docker uses as its storage driver.
Looks like the solution is to install the
linux-image-extra-virtual
, as concluded from moby/moby#10859 and other sources, and reboot the node. After reboot the node got back to theReady
state.Moreover, I guess an upgrade to a newer Docker version, using the official install command
curl -sSL https://get.docker.com/ | sh
would solve the problem as the most recent install script already installslinux-image-extra-virtual
as per moby/moby#10860, although I'm not sure how Docker was installed on the k8s node, so I don't know if it's safe to upgrade it this way.So I have a number of questions:
Note that I haven't done anything unusual, just installed fresh k8s on AWS using the default Ubuntu release. Hence I believe it's important to fix it somehow in k8s.
Possible duplicate: #9779.
The text was updated successfully, but these errors were encountered: