Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

On the evaluation under zero/few-shot setting with ITM+MLM #3

Open
matthewdm0816 opened this issue Aug 2, 2022 · 1 comment
Open

Comments

@matthewdm0816
Copy link

Hi,
Thanks for the open-source code provided in this repository for the paper, and I'm very interested in the zero/few-shot ability as shown in your great work, but as I investigate to evaluate the model under zero-shot MLM+ITM setting in run_gqa_prompt_zero_few.py, I found that the ITM head (i.e. model.cls_ans.seq_relationship) is newly initialized (and of different shape from original pretrained ITM head) and actually not used in the evaluation. Am I wrong on the usage of this code or is ITM head not participating in the evaluation?
Thanks a lot for your time and attention.

@LYuhang
Copy link

LYuhang commented Dec 28, 2022

@matthewdm0816 Thanks for your attention. Actually, we have initialized the cls_ans.seq_relationship in the code (refer to

def init_weights_with_pretrained_parameters(self, config):
)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants