Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Question about pretraining #12

Closed
PipiZong opened this issue Nov 30, 2021 · 1 comment
Closed

Question about pretraining #12

PipiZong opened this issue Nov 30, 2021 · 1 comment

Comments

@PipiZong
Copy link

Thanks for sharing your great work.

Just wondering which model do you select for the downstream task since you save pertaining models every 40 epochs.

Thank you.

@BestJuly
Copy link
Owner

Thank you for your interest.

For the pretraining part, I am just using the last epoch.
I have tried using different checkpoints for downstream task, the differences are limited, and there exist a better checkpoint. However, that is tricky and based on the testing performance. Therefore, only the last checkpoint will be used.

@BestJuly BestJuly closed this as completed Nov 8, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants