Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Did the number of parameters take into account the parameters in the tunable classification head? #16

Closed
ShengdingHu opened this issue Dec 30, 2021 · 1 comment

Comments

@ShengdingHu
Copy link

Thanks for releasing the code! I noticed that the reporting number of parameters of Lora module is 0.3 M for roberta-base. After experiments, I found that there are 0.5M parameters tunable in the sequence classification head (but it's the same for all baselines, so I am not arguing about the fairness). I wonder was I correct about the setting? Did the performances in the paper were the ones that also tune a classification head for classification tasks?

@ShengdingHu ShengdingHu changed the title Did the number of parameter take into account the parameters in the tunable language model head? Did the number of parameters take into account the parameters in the tunable classification head? Dec 30, 2021
@edwardjhu
Copy link
Collaborator

Hi Shengding,

You are right. We didn't include the classification head when counting parameters for all baselines, even though they were trained (otherwise they would be random). We'll add a note in our manuscript stating this explicitly.

P.S. My apology for the delayed response. I left Microsoft late last year and stopped receiving notifications.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants