Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Experiment with constant learning rate with transformers #24

Closed
cwmeijer opened this issue Nov 3, 2020 · 3 comments
Closed

Experiment with constant learning rate with transformers #24

cwmeijer opened this issue Nov 3, 2020 · 3 comments
Milestone

Comments

@cwmeijer
Copy link
Contributor

cwmeijer commented Nov 3, 2020

Do a couple of runs with constant learning, varied systematically between runs, and find out if we can find some learning rate that consistently reduces loss. After that we can experiment with more complicated schemes.

@cwmeijer cwmeijer added this to the Transformers milestone Nov 3, 2020
@egpbos
Copy link
Collaborator

egpbos commented Nov 13, 2020

I started a run, effortless-cosmos-19, with the cyclic scheduler with maximum and minimum set to values that may be more sane for this model (based on this article). So far, it is at least outperforming all previous runs. Will let it run until Monday 8 a.m. (according to my estimation) and then we'll see how it went.

Already, now, one interesting feature is the correlation of the periodic pattern in the loss with the learning rate. This might suggest that, as also mentioned in the article, some kind of decreasing maximum might be helpful to press down the big (seemingly useless) peaks in the loss and make it converge even faster.

@egpbos
Copy link
Collaborator

egpbos commented Nov 13, 2020

New run dainty-dawn-20 also logs validation loss on wandb.

@egpbos
Copy link
Collaborator

egpbos commented Nov 16, 2020

I think we can close this, since the training loss is very effectively reduced now. Next problem is validation loss, issue #30.

@egpbos egpbos closed this as completed Nov 16, 2020
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants