Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

The MetaQA dataset #125

Open
LinxiCai opened this issue Nov 10, 2022 · 2 comments
Open

The MetaQA dataset #125

LinxiCai opened this issue Nov 10, 2022 · 2 comments

Comments

@LinxiCai
Copy link

when I read the train.txt, valid.txt and test.txt in data/MetaQA, I found the triples in test.txt are included in train.txt, could you explain why should this happen?

@apoorvumang
Copy link
Collaborator

Hi LinxiCai, thanks for your interest.

We studied MetaQA for the QA task, not KG completion task. We want to pretrain on the whole KG (or 50% KG depending on setting) and then finetune for QA. test.txt and valid.txt triples exist just for compatibility with KGE implementations, which require separate validation and test triples. So we simply copied triples from train.txt to test.txt to maintain compatibility.

@LinxiCai
Copy link
Author

OK,thanks a lot !! I understand. By the way, I had another question, when I train embedding for metaQA triples, if I don't use dropout or batch_normalization, will there be overfitting? or can you share your training arguments when you get your MetaQA embedding?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants