-
Notifications
You must be signed in to change notification settings - Fork 24.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Sharding based on shard size or index size #16828
Comments
A few of the Elasticsearch committers have been kicking this idea around for a while. I wasn't the one who originated it but I'm online now so I'll describe it: Right now if you have time based data like logs then we (Elastic the company, Elasticsearch committers, helpful people on discuss.elastic.co) recommend you go with an index per X time period. It makes you chose a few things:
So we have a plan to make all of these choices simpler! The basics go like this: create an index with enough shards to get maximal write performance. Something like Now, you may ask lots of interesting questions. Like:
I'll add to this list/change it as I think of more things. Does that cover it? |
Almost certainly we'll do this as a part of the index's configuration, probably near where you'd set number_of_shards. |
This is closed by the rollover and shrink APIs. Closing. |
Describe the feature:
It would be nice to be able to specify a target shard size in ES config file, and when a shard approaches that size, ES automatically create a new "shard" for that index. Our use case is a dynamic log size for our indices affected by organic traffic pattern changes and traffic shifts from one datacenter to a different one. These changes result in shards with undesired size, even up to 150 Gb, being relocated frequently, and causing instability to the entire cluster itself. Tuning the shard size is not an straight forward process and there may not be a single solution to the equation.
The text was updated successfully, but these errors were encountered: