Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

DIsk based allocator should take into account the amount of disk space which can be reused #10103

Closed
maf23 opened this issue Mar 16, 2015 · 8 comments
Labels
:Distributed/Allocation All issues relating to the decision making around placing a shard (both master logic & on the nodes) stalled

Comments

@maf23
Copy link

maf23 commented Mar 16, 2015

Assume we have a node where the disk usage is above the low watermark. If we restart this node then no shards will be allocated on it since the disk usage is above the low watermark.

It would be better if the disk based allocator could realize that it can throw away or at least reuse the space which is currently used for outdated shard copies on the node.

@clintongormley
Copy link

@dakrone any thoughts about this?

@dakrone
Copy link
Member

dakrone commented May 8, 2015

@clintongormley hmm.. to implement this we would have to send a request for each shard to all nodes to see what files and checksums they have in common. This would be quite heavy, especially for an AllocationDecider (because it has to run in tight loops).

@clintongormley
Copy link

@dakrone thanks for commenting, that sounds... heavy. Perhaps once sequence IDs (#10708) are in, it'll be more feasible.

Marking as stalled.

@clintongormley
Copy link

@dakrone @bleskes I think this has been resolved already, no?

@dakrone
Copy link
Member

dakrone commented Dec 9, 2015

@clintongormley no, it hasn't be resolved still, I'm not sure if the seqids work is far enough along to work on this, maybe @bleskes knows?

@bleskes
Copy link
Contributor

bleskes commented Dec 9, 2015

I don't think we can expect this in the first iteration. The problem is very similar to the current file based copy - we don't know how much will be reused (or if seq no base recovery will succeed) until we start the work. The master needs to know this in advance somehow (as Lee already said).

@lcawl lcawl added :Distributed/Distributed A catch all label for anything in the Distributed Area. If you aren't sure, use this one. and removed :Allocation labels Feb 13, 2018
@DaveCTurner DaveCTurner added the :Distributed/Allocation All issues relating to the decision making around placing a shard (both master logic & on the nodes) label Mar 15, 2018
@elasticmachine
Copy link
Collaborator

Pinging @elastic/es-distributed

@DaveCTurner DaveCTurner removed the :Distributed/Distributed A catch all label for anything in the Distributed Area. If you aren't sure, use this one. label Mar 15, 2018
@DaveCTurner
Copy link
Contributor

As I understand it, the situation described here is a transient problem around a node restart in the sense that the cluster sorts itself out eventually once it works out what data it no longer needs. As such, since its opening we have not seen enough feedback that it is something we should work on. We prefer to close this issue as a clear indication that we are not going to work on this at this time. We are always open to reconsidering this in the future based on compelling feedback; despite this issue being closed please feel free to leave feedback (including +1s).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
:Distributed/Allocation All issues relating to the decision making around placing a shard (both master logic & on the nodes) stalled
Projects
None yet
Development

No branches or pull requests

7 participants