Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

layernorm should be set to trainable or not ? #98

Closed
rabeehkarimimahabadi opened this issue Nov 29, 2020 · 2 comments
Closed

layernorm should be set to trainable or not ? #98

rabeehkarimimahabadi opened this issue Nov 29, 2020 · 2 comments
Labels
question Further information is requested

Comments

@rabeehkarimimahabadi
Copy link

Hi,
I see in some implementations they set layer_norm as require_grad=True, could you tell me if all layer norms of the model needs to be set to require_grad=True, or only the ones inside adapter layer needs this condition?
thanks.

@rabeehkarimimahabadi rabeehkarimimahabadi added the bug Something isn't working label Nov 29, 2020
@arueckle arueckle added question Further information is requested and removed bug Something isn't working labels Nov 29, 2020
@arueckle
Copy link
Member

See figure 3 in the adapterhub paper, everything outside the pink box is fixed/frozen during adapter training

@rabeehkarimimahabadi
Copy link
Author

rabeehkarimimahabadi commented Nov 29, 2020 via email

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

2 participants