-
Notifications
You must be signed in to change notification settings - Fork 23
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Use tensor_flow_addons #170
Conversation
ndiamant
commented
Mar 16, 2020
- learning rate finding, and find then train recipes added
- learning rate schedules added (triangular, triangular2)
- RAdam dependency removed
- lookahead removed (could be easily readded with tfa)
- More normalizations added, though inaccessible from the command line now, because untested
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Awesome! Some small nitpicks inline. Bigger question is we still use ReduceLROnPlateau
by default during training, does that still make sense with the LR finding done here? while we're at it should we make that command line toggle-able
@lucidtronix |
Reminder to push docker with tfa before merge |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks good! If tests pass push docker (GPU & CPU) images and merge at will!
Tests pass!
* rectified adam * removed lookahead, because it is unused and could be implemented with tfa * added more activation options * more normalization schemes, learning rate schedules * added learning rate finder * bug fixes * delta for best lr picking in its own function * smoothing finds better lrs * small fix * plot_metrics handles learning rate schedulers * learning rate schedule works with _find_lr * reorganization * bug fix Co-authored-by: ndiamant <ndiamant@broadinstitute.org>
* rectified adam * removed lookahead, because it is unused and could be implemented with tfa * added more activation options * more normalization schemes, learning rate schedules * added learning rate finder * bug fixes * delta for best lr picking in its own function * smoothing finds better lrs * small fix * plot_metrics handles learning rate schedulers * learning rate schedule works with _find_lr * reorganization * bug fix Co-authored-by: ndiamant <ndiamant@broadinstitute.org>
* rectified adam * removed lookahead, because it is unused and could be implemented with tfa * added more activation options * more normalization schemes, learning rate schedules * added learning rate finder * bug fixes * delta for best lr picking in its own function * smoothing finds better lrs * small fix * plot_metrics handles learning rate schedulers * learning rate schedule works with _find_lr * reorganization * bug fix Co-authored-by: ndiamant <ndiamant@broadinstitute.org>
* rectified adam * removed lookahead, because it is unused and could be implemented with tfa * added more activation options * more normalization schemes, learning rate schedules * added learning rate finder * bug fixes * delta for best lr picking in its own function * smoothing finds better lrs * small fix * plot_metrics handles learning rate schedulers * learning rate schedule works with _find_lr * reorganization * bug fix Co-authored-by: ndiamant <ndiamant@broadinstitute.org>