Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

The task-specific weights for h_phi are not adapted in the inner loop #2

Closed
xuegsh opened this issue Nov 30, 2020 · 1 comment
Closed

Comments

@xuegsh
Copy link

xuegsh commented Nov 30, 2020

Hello, sir. I run the code in the repository and got some questions about it :

  1. I use the script to run finetuning on the CoNLL task and I notice that the weights of the 2-layer MLP (which is denoted as h_phi in the paper) do not change as the training goes on. And I found that the reason is that the per-layer learning rates for h_phi are initialized to 0 and are set to untrainable as follows:
    image

This means that the task-specific weights for h_phi are not adapted in the inner loop. It is inconsistent with what the paper says.

  1. The learning rates for the 2-layer MLP (which is denoted as g_psi in the paper) seems redundant as they haven't been used in the adaption phase.
@theTB
Copy link
Contributor

theTB commented Dec 2, 2020

Hi. As you can see from the code snippet that you shared, these are treated as warp layers and hence they are not adapted during fine-tuning. So what you are seeing is the intended behavior.

@theTB theTB closed this as completed Dec 2, 2020
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants