-
Notifications
You must be signed in to change notification settings - Fork 24.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Node leaving the cluster does not trigger 'recover', unassigned shards are not re-assigned #15003
Comments
A manual reroute should not be needed. This sounds like a bug. Can you share information about the size of the cluster and the settings used, especially those affecting shard allocation (https://www.elastic.co/guide/en/elasticsearch/reference/2.1/index-modules-allocation.html)? |
Cluster has been fluctuating between 6 and 3 nodes, to test the scaling. It holds 12 indexes, with 3 shards and 1 replica. It holds 75m documents, in total 100G of data. (Kind of augmented logstash setup ingesting a bit more than 3 documents per second, sustained.) It runs on EC2. Normally we use 3 t2.medium instances (it is staging) but we have been rotating up to m4.10xlarge because we had to restore a month. elasticsearch.json:
|
Just double checking - this is not by any chance a case of delayed assignment? - see https://www.elastic.co/guide/en/elasticsearch/reference/current/delayed-allocation.html#delayed-allocation |
(hi boaz :) ) no. i tried to put that in the config, to force an automatic explicit trigger :) (couldn't find how to do that.) |
I have a similar problem in a 8 nodes cluster, 7 data nodes and 1 non-data. This is using ES 2.0 also. |
Somehow it allocated the shards after a few hours, although delayed_unassigned_shards showed 0 the whole time. |
I tried to to reproduce this and couldn't - I started a 4 nodes cluster, added 12 indices and your settings. Shutting down a node (kill -9) resulted in the expected 60s wait for it to come back, followed by allocation of the missing shards and back to green. @truthtrap can you supply more info/work on a small and reliable reproduction? |
@bleskes thanks for trying. can you try an 'orderly' shutdown with the init.d script that is part of the package (rpm we use)? i'll try to set up a completely standard cluster, populate with some indices, and try to reproduce. |
@truthtrap is this something you're still seeing on 2.2? |
@clinton didn't try yet. will upgrade to 2.2.0, and let you know... On Sun, Feb 14, 2016 at 6:54 PM, Clinton Gormley notifications@github.com
|
@clinton looks good so far. upgraded (and rotated) our staging cluster. it (our staging cluster has 125G, 14 indices, each 3 shards with 1 replica.) On Mon, Feb 15, 2016 at 11:13 AM, Jurg van Vliet jurg@truthtrap.com wrote:
|
@clinton rotating production elasticsearch cluster was ok as well. no thanks!! On Wed, Feb 17, 2016 at 7:58 AM, Jurg van Vliet jurg@truthtrap.com wrote:
|
thanks @truthtrap - closing |
Dear people,
After upgrading to ES 2.0.0 I noticed that sometimes shards stay unassigned after a node leaves the cluster. We have been looking at all the usual suspects like
but, the only thing that we found working in this situation was a manual reroute or enabling the allocation (which is already enabled). Later it became clear that it only happens on stable clusters. If there are relocations going on the unassigned shards are assigned fine.
Yesterday we applied the 2.1.0 update. But we still have the same issue.
As far as we can see this is an issue (for us certainly). If it is a feature my apologies for abusing this area, and I will move the discussion to the forums.
Groet,
Jurg.
The text was updated successfully, but these errors were encountered: