-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Cluster deployment hangs at "Failed to connect events watcher: Get https://mycluster-api.openshift.testing:6443/apikube-system/events? #721
Comments
Sometimes the |
I did clean up the state (rm -rf and virt-cleanup.sh) and run the installer in a new dir.
|
I somehow missed that the issue is with etcd certs. In order to reach the event watcher, you need to have a reasonably happy etcd cluster. If you run that health check after seeing this TLS problem, is your etcd cluster still reporting itself as healthy? And when you connect to one of the etcd nodes on 2379, is it giving you a cert signed by the generated etcd CA? |
It's healthy. |
on the hypervisor dnsmasq, there's no a record for etcd.kube-system.svc
|
can you paste output for
is a kubernetes services for aggregated servers to talk to etcd cluster. you can check the state of the services using |
the api pods are down. That's my original issue. |
This should be resolved by now. |
Version
Platform (aws|libvirt|openstack):
Libvirt
What happened?
The installation stuck at the check of the api.
level=warning msg="Failed to connect events watcher: Get https://mycluster-api.openshift.testing:6443/api/v1/namespaces/kube-system/events?resourceVersion=2176&watch=true: dial tcp 192.168.126.10:6443: connect: connection refused"
The etcd is running with some issue regarding the tls certificate:
On the pod of api-opneshift and api-kube-openshift are dead:
logs of openshift-apiserver
However on the openshift-kube-api I see that it was not able to resolve the etcd.kube-system.svc.
(I don't know if it's related).
What you expected to happen?
finish the cluster installation process: the api is up, and the new worker node was created.
How to reproduce it (as minimally and precisely as possible)?
create-cluster mycluster
The text was updated successfully, but these errors were encountered: