Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Resetting of high scores, scheduler and optimizer for fine-tuning/domain adaptation #75

Merged
merged 2 commits into from Nov 25, 2019

Conversation

juliakreutzer
Copy link
Collaborator

In the case of fine-tuning or domain adaptation, one might want to overwrite the existing scheduler or optimizer or the previously tracked high score.
This is specifically relevant when fine-tuning with a different metric or a data where the validation score is in a different range than the loaded model (e.g. pre-training on BLEU and fine-tuning on PPL would result in never storing a new checkpoint). When the data is much smaller than the previous training data, one might want to reduce the patience.

We can also think about modeling this as "load_x_from_ckpt" rather than "reset_x", but for now, it seems that in the default case we want to load everything.

@joeynmt joeynmt requested review from dwhitena and removed request for bastings November 25, 2019 20:36
Copy link

@dwhitena dwhitena left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This works for me. When the flags are set, I can see the checkpoints being saved in the filesystem and logged properly. Thanks for updating this!

@juliakreutzer
Copy link
Collaborator Author

Awesome, thanks for confirming this so quickly @dwhitena! And an official welcome as collaborator! 🎉

@juliakreutzer juliakreutzer merged commit 1f51b40 into master Nov 25, 2019
@juliakreutzer juliakreutzer deleted the reset_loading branch November 25, 2019 23:32
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

2 participants