Navigation Menu

Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[ML] allow autoscaling to work when vertical scaling is possible #84242

Conversation

benwtrent
Copy link
Member

When an NLP model is deployed, or a DFA/Anomaly job is assigned, we have historically relied only on the xpack.ml.max_lazy_ml_nodes to determine if scaling is possible. But, in certain scenarios, it may be that scaling is available when xpack.ml.max_lazy_ml_nodes is fully satisfied.

xpack.ml.max_ml_node_size is now checked to see if the current ML nodes exceed this size. If not, we assume vertical scaling is possible and allow the tasks to be created.

closes #84198

@benwtrent benwtrent added >bug :ml Machine learning v8.2.0 labels Feb 22, 2022
@elasticmachine elasticmachine added the Team:ML Meta label for the ML team label Feb 22, 2022
@elasticmachine
Copy link
Collaborator

Pinging @elastic/ml-core (Team:ML)

@elasticsearchmachine
Copy link
Collaborator

Hi @benwtrent, I've created a changelog YAML for you.

Copy link
Contributor

@droberts195 droberts195 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

@droberts195
Copy link
Contributor

This fix should be backported to 7.17 and 8.1, otherwise lots of 7.17 users could be affected over the coming years.

@benwtrent benwtrent added v7.17.1 v8.1.1 auto-backport-and-merge Automatically create backport pull requests and merge when ready labels Feb 22, 2022
@benwtrent benwtrent added auto-backport Automatically create backport pull requests when merged and removed auto-backport-and-merge Automatically create backport pull requests and merge when ready labels Feb 23, 2022
@benwtrent benwtrent merged commit 28758b0 into elastic:master Feb 23, 2022
@benwtrent benwtrent deleted the feature/ml-fix-lazy-assignment-when-virt-scaling-is-possible branch February 23, 2022 12:54
benwtrent added a commit to benwtrent/elasticsearch that referenced this pull request Feb 23, 2022
…stic#84242)

When an NLP model is deployed, or a DFA/Anomaly job is assigned, we have historically relied only on the xpack.ml.max_lazy_ml_nodes to determine if scaling is possible. But, in certain scenarios, it may be that scaling is available when xpack.ml.max_lazy_ml_nodes is fully satisfied.

xpack.ml.max_ml_node_size is now checked to see if the current ML nodes exceed this size. If not, we assume vertical scaling is possible and allow the tasks to be created.

closes elastic#84198
@elasticsearchmachine
Copy link
Collaborator

💔 Backport failed

Status Branch Result
8.1
7.17 Commit could not be cherrypicked due to conflicts

You can use sqren/backport to manually backport by running backport --upstream elastic/elasticsearch --pr 84242

benwtrent added a commit to benwtrent/elasticsearch that referenced this pull request Feb 23, 2022
…stic#84242)

When an NLP model is deployed, or a DFA/Anomaly job is assigned, we have historically relied only on the xpack.ml.max_lazy_ml_nodes to determine if scaling is possible. But, in certain scenarios, it may be that scaling is available when xpack.ml.max_lazy_ml_nodes is fully satisfied.

xpack.ml.max_ml_node_size is now checked to see if the current ML nodes exceed this size. If not, we assume vertical scaling is possible and allow the tasks to be created.

closes elastic#84198
elasticsearchmachine pushed a commit that referenced this pull request Feb 23, 2022
#84242) (#84280)

* [ML] allow autoscaling to work when vertical scaling is possible (#84242)

When an NLP model is deployed, or a DFA/Anomaly job is assigned, we have historically relied only on the xpack.ml.max_lazy_ml_nodes to determine if scaling is possible. But, in certain scenarios, it may be that scaling is available when xpack.ml.max_lazy_ml_nodes is fully satisfied.

xpack.ml.max_ml_node_size is now checked to see if the current ML nodes exceed this size. If not, we assume vertical scaling is possible and allow the tasks to be created.

closes #84198

* fixing for backport
probakowski pushed a commit to probakowski/elasticsearch that referenced this pull request Feb 23, 2022
…stic#84242)

When an NLP model is deployed, or a DFA/Anomaly job is assigned, we have historically relied only on the xpack.ml.max_lazy_ml_nodes to determine if scaling is possible. But, in certain scenarios, it may be that scaling is available when xpack.ml.max_lazy_ml_nodes is fully satisfied.

xpack.ml.max_ml_node_size is now checked to see if the current ML nodes exceed this size. If not, we assume vertical scaling is possible and allow the tasks to be created.

closes elastic#84198
benwtrent added a commit that referenced this pull request Feb 28, 2022
) (#84286)

When an NLP model is deployed, or a DFA/Anomaly job is assigned, we have historically relied only on the xpack.ml.max_lazy_ml_nodes to determine if scaling is possible. But, in certain scenarios, it may be that scaling is available when xpack.ml.max_lazy_ml_nodes is fully satisfied.

xpack.ml.max_ml_node_size is now checked to see if the current ML nodes exceed this size. If not, we assume vertical scaling is possible and allow the tasks to be created.

closes #84198
nik9000 pushed a commit to nik9000/elasticsearch that referenced this pull request Feb 28, 2022
…stic#84242)

When an NLP model is deployed, or a DFA/Anomaly job is assigned, we have historically relied only on the xpack.ml.max_lazy_ml_nodes to determine if scaling is possible. But, in certain scenarios, it may be that scaling is available when xpack.ml.max_lazy_ml_nodes is fully satisfied.

xpack.ml.max_ml_node_size is now checked to see if the current ML nodes exceed this size. If not, we assume vertical scaling is possible and allow the tasks to be created.

closes elastic#84198
tlrx pushed a commit to tlrx/elasticsearch that referenced this pull request Mar 3, 2022
…stic#84242)

When an NLP model is deployed, or a DFA/Anomaly job is assigned, we have historically relied only on the xpack.ml.max_lazy_ml_nodes to determine if scaling is possible. But, in certain scenarios, it may be that scaling is available when xpack.ml.max_lazy_ml_nodes is fully satisfied.

xpack.ml.max_ml_node_size is now checked to see if the current ML nodes exceed this size. If not, we assume vertical scaling is possible and allow the tasks to be created.

closes elastic#84198
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
auto-backport Automatically create backport pull requests when merged >bug :ml Machine learning Team:ML Meta label for the ML team v7.17.2 v8.1.1 v8.2.0
Projects
None yet
Development

Successfully merging this pull request may close these issues.

ML autoscaling decider is not taking into account that an existing node can be upscaled
5 participants