Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Do not persist entire AutoMLState in Searcher #870

Merged

Conversation

Yard1
Copy link
Collaborator

@Yard1 Yard1 commented Jan 4, 2023

Signed-off-by: Antoni Baum antoni.baum@protonmail.com

Why are these changes needed?

Currently, the entire AutoMLState is persisted in Searcher as a part of the config_constraints. As the State object contains the dataframe used, this means a large amount of data will be persisted to disk as a part of automatic Tune Experiment checkpointing, leading to large disk usage and high performance overhead, causing various bottlenecks.

This PR changes the size function used in config_constraints to only consider the learner_classes attribute of AutoMLState which removes the issue of too much data being saved.

Related issue number

Closes #866

Checks

Signed-off-by: Antoni Baum <antoni.baum@protonmail.com>
Signed-off-by: Antoni Baum <antoni.baum@protonmail.com>
Copy link
Collaborator

@sonichi sonichi left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Looks good to me. Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Low cpu usage when parallel tuning
3 participants