Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

about the ckp epoch #2

Closed
Vijayue opened this issue Oct 19, 2021 · 2 comments
Closed

about the ckp epoch #2

Vijayue opened this issue Oct 19, 2021 · 2 comments

Comments

@Vijayue
Copy link

Vijayue commented Oct 19, 2021

Thanks a lot for your code and pre-trained model.

Now I want to continue training on your pretrained model, after loading the pretrained model, the epoch begins from the 12400, but the ckp name is 00000100-ckp.pth.tar, which means the ckp was generated after the 100 epochs? Do you have any idea about the issue? Thank you!

@zhengkw18
Copy link
Owner

I'm sorry this mismatch may confuse you. This is caused by the change of 'num_repeats' in 'train.py'. When I train the model, it is set to 1, so each epoch takes only a few minutes to facilitate faster debugging. When I release the code, it is set to 100, which is consistent with FOMM. In the current config, the checkpoint is roughly at the progress of 124 epoch.

@Vijayue
Copy link
Author

Vijayue commented Oct 19, 2021

OK! Thank you for your explanation, I got it!

@Vijayue Vijayue closed this as completed Oct 19, 2021
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants