New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Master node will not restart on AWS #16188
Comments
It's relevant that the cloud init scripts passed in through the kube-up.sh only run on first boot, not subsequent ones. |
Apparently, the LVM for kubernetes is not getting created correctly, so on a restart it can't find the volume. This is on 1.2 alpha, it may be different on 1.0.6.
|
Some relevant debug info:
|
In stepping through the code in format-disks.sh, the error starts here (on non-wheezy versions of Ubuntu... it looks like wheezy is fine):
Then checking the details:
|
As a test, I ran the wheezy version of lvcreate (no thin provisioning) manually and restarted. It worked, and master came back up correctly. |
+1. We have experienced the issue on AWS too. The nodes (Ubuntu) did not come back after reboot. |
+1 |
Apparently, the metadata is written into the pool and thin provisioning will not work with 100%FREE: |
/sub @quinton-hoole |
Will this patch be applied to the 1.1 release branch? |
just met this issue on v1.1.3 (kubectl config was empty, probably lost data on attached ebs storage, wasn't able to re-gain control and recover with etcd, couldn't find more info about how to re-configure master back to existing cluster on aws) kubectl get nodes :: from kubelet log :: |
this is a major deal here for production use.
|
This fix has been pulled into 1.2, but is not in 1.1. We run 1.1 but during cluster up, we copy master's cluster/aws/templates/format-disks.sh |
If you create a cluster in AWS using kube-up.sh it uses cloud init to set the box up. On a restart of the node the script fails. This is true for 1.0.6 and 1.2.0-alpha.2 From the system log:
The text was updated successfully, but these errors were encountered: