Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Nomad didn't deregister services in Consul #810

Closed
adrianlop opened this issue Feb 17, 2016 · 6 comments
Closed

Nomad didn't deregister services in Consul #810

adrianlop opened this issue Feb 17, 2016 · 6 comments

Comments

@adrianlop
Copy link
Contributor

Hi there,

I have Consul 0.6.3 and Nomad 0.2.3 agents running as a system service.
When I rebooted some machines, the services running on these machines reallocated to other available machines. That worked like a charm.
The problem was that the services were still registered (but not running, I double-checked that) on the old machines, and in the new ones.
I checked Nomad logs in the old machines and there's no reference about the allocID (even after restarting the Nomad agent too).
So my guess is that when the reboot happened, Nomad cluster saw that the machines were unreachable, so it reallocated the services, but when they came up again, the Nomad agents didn't send a query to Consul to deregister the services.

Can you please help? is this fix in Nomad 0.3 maybe?

@diptanu
Copy link
Contributor

diptanu commented Feb 17, 2016

@adrianlop Can you share the Nomad client logs of the machines where the services were supposed to be not registered?

@adrianlop
Copy link
Contributor Author

Ok, I'll rerun the reboots. do you prefer them on DEBUG?

@diptanu
Copy link
Contributor

diptanu commented Feb 17, 2016

Yes, please share the debug logs.

Sent from my iPhone

On Feb 17, 2016, at 9:21 AM, Adrián López notifications@github.com wrote:

Ok, I'll rerun the reboots. do you prefer them on DEBUG?


Reply to this email directly or view it on GitHub.

@adrianlop
Copy link
Contributor Author

now that you released v0.3, I'm going to upgrade the whole cluster first, then I'll try to reproduce this again. thanks!

@adrianlop
Copy link
Contributor Author

hi again,
after upgrading to Nomad 0.3, I rebooted several servers and everything went OK.

I think what happened is:

So I'm going to close the issue for now, if I reboot machines some other times, and there are no zombie Dockers, then I'll reopen and send you debug logs.

Thanks @diptanu !

@github-actions
Copy link

I'm going to lock this issue because it has been closed for 120 days ⏳. This helps our maintainers find and focus on the active issues.
If you have found a problem that seems similar to this, please open a new issue and complete the issue template so we can capture all the details necessary to investigate further.

@github-actions github-actions bot locked as resolved and limited conversation to collaborators Dec 25, 2022
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Projects
None yet
Development

No branches or pull requests

2 participants