Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

are predecessor modules weights frozen or not #9

Closed
TobiasLee opened this issue Aug 12, 2020 · 4 comments
Closed

are predecessor modules weights frozen or not #9

TobiasLee opened this issue Aug 12, 2020 · 4 comments

Comments

@TobiasLee
Copy link

Hi, thanks for your great work.
According to the paper description, the predecessor module weights are frozen after fine-tuned on the task data( including embedding & output classifier).
The code, however, if my understanding is correct, the fine-tuned predecessor weights are not frozen, instead, the loss can BP to the corresponding parameters.
So, which pattern is supposed to be correct? Thanks in advance.

@JetRunner
Copy link
Owner

They are frozen. Please check the optimizer. Thanks!

@TobiasLee
Copy link
Author

Thanks for the notification.

@Ca0L
Copy link

Ca0L commented May 20, 2021

Sorry, I don't see how the optimizer freeze predecessor modules weights. Could you please explain it? Thanks.

@JetRunner
Copy link
Owner

Sorry, I don't see how the optimizer freeze predecessor modules weights. Could you please explain it? Thanks.

They are not in the optimized parameters of the optimized. PyTorch optimizers can only change the parameters passed to them.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants