Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

ZK connectivity failure with multiple watchers leads to permanent failure #92

Closed
minkovich opened this issue Apr 18, 2017 · 3 comments
Closed

Comments

@minkovich
Copy link

Setup:
6 nerve service watchers on the same instance connected to the same ZK pool

How to reproduce:

  1. The instance has a problem connecting to ZK
  2. Nerve ->
    Nerve::Nerve: nerve: watcher service1 not alive; reaping and relaunching
    Nerve::ServiceWatcher: nerve: stopping service watch service1
    Nerve::Nerve: nerve: could not reap service1, got #<Zookeeper::Exceptions::NotConnected: Zookeeper::Exceptions::NotConnected>
  3. This continues in a loop for each service watcher until nerve is restarted.

Actual problem:
The problem is that in start() in zookeeper.rb there are no checks to see if the ZK connection is alive before re-using in.

@jolynch
Copy link
Collaborator

jolynch commented Apr 19, 2017

@minkovich what is your desired behavior here? I suppose that we would like it if Nerve threw out the bad cached connection and tried again?

If the cluster is just not reachable this would lead to a similar infinite retry loop, but perhaps crash-recover is sufficient here?

@minkovich
Copy link
Author

@jolynch The efficient solution would be for nerve to throw away the bad connection, but honestly in this situation a crash recovery would also be equivalent since connectivity was already lost.

@panchr
Copy link
Contributor

panchr commented Feb 13, 2020

Closed because this was fixed in #113.

@panchr panchr closed this as completed Feb 13, 2020
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

3 participants