-
Notifications
You must be signed in to change notification settings - Fork 89
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Readiness probe failed: node id file does not exist - pod is not yet bootstrapped #177
Comments
Hi, What does |
It returns this:
|
Thank you. What is the output of the 2 following commands: |
Output of kubectl describe pvc :
Output of kubectl describe pv :
|
Hi, I would suggest running the log_collector.py script from this project in order to generate a diagnostic package and open a Support Ticket with Redis Labs to upload the package so that we can analyze it. Laurent. |
Thanks for the help. I have opened a ticket with Redis Labs for the same. |
Hello @pinkeshr, |
Hi Paul, This is a very generic message that is always displayed when the pod (node) is not bootstrapped. There can be dozens of reasons for it, so just like above, I'd suggest running the log_collector.py script from this project in order to generate a diagnostic package and open a Support Ticket with Redis to upload the package so that we can understand what is causing this on your cluster and help you with it. Cheers, Laurent. |
ITNOA I have the same issue
My cluster.yaml file like below
I run
What is my problem? |
@laurentdroin I think my problem from |
Hi Soorosh, I think the problem for you is the resources. 1 GB of memory is definitely not enough and the first pod will never be able to create the cluster. Can you increase the memory to at least 3 GB and let me know if this helped? Laurent. |
I add some memory to our cluster, and then increase memory of environment, so my problem is resolved. @laurentdroin thanks for helping |
@laurentdroin Hi again, After I increase my memory and resolve my previous problem, and all of things, work properly. I try to power off all nodes. after some day and turn on again my system, I see below
as you can see
and bootstrapper log like below
As you can see my pods have sufficient memory, How to resolve my problem? |
related to #214 |
Hi @soroshsabz, Yes, as I explained in the other issue you opened (I didn't see you opened this new one), this is expected. If your cluster has lost quorum and is therefore no longer working, you would need to recover it using this procedure: https://docs.redis.com/latest/kubernetes/re-clusters/cluster-recovery/ I hope this helps. Laurent. |
After remove cluster, I can create new and healthy cluster Thanks to @laurentdroin |
Hello, got the same problem, @laurentdroin i understand that we need to maintain the cluster at all times to get that working fine. But since we are using an Enterprise Operator (and we pay for it) can you make the magic happens and automatically resolve this problem in case of emergency reboot of our K8S nodes. If we are moving from standalone redis or redis/sentinel or even google memorystore to redis entreprise it's for have more: Actually we cannot move forward using that product in this state, we do not ask the moon... just something that's works |
@cschockaert thank you for the feedback! |
Hello, i'm in touch with redis team (@fcerbelle), (not directly the support). Actually, we are using the preemptible (https://cloud.google.com/kubernetes-engine/docs/how-to/preemptible-vms) feature of GKE (node can be killed anytime after 24H of life ; but cost us 10x less) |
Hi, I am trying to set up a Redis cluster on gke using operator but it fails with error:
Readiness probe failed: node id file does not exist - pod is not yet bootstrapped.
I have a gke cluster up and running with
node-pool of 6 nodes
andmachine type n1-standard-8.
Steps to reproduce:
kubectl apply -f bundle.yaml
kubectl apply -f rec.yaml
. It fails.These are the event logs found from command
kubectl describe pod redis-enterprise-0
My rec.yaml file looks like this :
I have tried with different CPU limits but still facing the same error.
Please let me know if I am doing something wrong here.
The text was updated successfully, but these errors were encountered: