Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Single-pass Self Parameter Tuning #1553

Draft
wants to merge 3 commits into
base: main
Choose a base branch
from
Draft

Single-pass Self Parameter Tuning #1553

wants to merge 3 commits into from

Conversation

smastelini
Copy link
Member

@smastelini smastelini commented May 31, 2024

This has been in the works for quite a long time, actually.

SSPT works by self-tuning the hyperparameters of models while keeping their parameters intact. For example, we might change some hyperparameters of linear models while keeping the weights learned so far intact. This is different from instantiating a new model from scratch with a set of candidate hyperparameter values.

@BrunoMVeloso is one of the authors and he is going to help me with the final touches.

The main things to do:

  • write documentation
  • remove numpy usage (it should be straightforward)
  • check the API: is the way hyperparameter exploration domains are supplied adequate?
  • verify and discuss compatibility

About the last item, the way SSPT is structured right now makes it inherit directly from base.Estimator. The good part is that you can use it to work with classifiers, regressors, clustering algorithms, and anomaly detectors (there is no need for a SSPTClassifier and so on, for example). The downside is that SSPT currently does not work with evaluate.progressive_val_score` and friends because the evaluators check internally whether the estimator is a classifier, regressor, and whatnot.

Saulo Martiello Mastelini added 2 commits May 31, 2024 10:49
@smastelini smastelini self-assigned this May 31, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant