-
-
Notifications
You must be signed in to change notification settings - Fork 2.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add better learning rate annealing scheduler + Learning rate finder #228
Comments
That sounds like a great idea - if fact we've been recently thinking a lot about better methods to find good parameters than a full parameter sweep. Would you be interested in testing/integrating such approaches? |
@kashif what do you think? |
@mhham @alanakbik Indeed I have a PR in preparation that adds weight decay versions of optimizers after which we can add the corresponding schedulers that not only change the learning rate, but also the weight decay factor or momentum |
Learning rate finder was just added for release-0.4 - we'll check out the cosine annealing in a later version. |
Two ideas that could allow improving training :
Add learning rate finder : as explained in here, and implemented here
Cosine Annealing with restart : https://github.com/roveo/pytorch/pull/1/commits/35891c46dbdec46e55a5fb725418c2880f631dac
The text was updated successfully, but these errors were encountered: