Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Implement in-training checkpoints for algos #7192

Closed
exalate-issue-sync bot opened this issue May 11, 2023 · 2 comments
Closed

Implement in-training checkpoints for algos #7192

exalate-issue-sync bot opened this issue May 11, 2023 · 2 comments
Assignees

Comments

@exalate-issue-sync
Copy link

h3. Edit: The original task has changed to implement in-training checkpoints

{quote}
Grid search has the ability to checkpoint models but each model is independent. If a user can extend a model during grid search, they can use learnings from a previous model to grow another. {quote}

{quote}For example, you can make a GBM grid with ntrees=[50, 70, 100]. The grid search can first build 50 trees, checkpoint, then extend that previous model by adding more trees (50+20 = 70 trees), and so on. We are assuming that other hyperparameters are fixed in the grid.

This will reduce the overhead of repeating the work of smaller models and can also allow users to run training on the independent sessions while saving learnings/checkpoints from previous runs.{quote}

@h2o-ops-ro
Copy link
Collaborator

JIRA Issue Details

Jira Issue: PUBDEV-8470
Assignee: Adam Valenta
Reporter: Neema Mashayekhi
State: Resolved
Fix Version: 3.38.0.1
Attachments: N/A
Development PRs: Available

@h2o-ops-ro
Copy link
Collaborator

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants