-
Notifications
You must be signed in to change notification settings - Fork 18
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Inference time when calulating MRR #8
Comments
Hi, thanks for your interest. evaluate.py does evaluation in the 'traditional' sense - it calculates log probabilities of all entity sequences (5M) using the Transformer model and then ranks these entities. It does not generate predictions, hence the long time. Please use the code in eval_accelerate.py for the sampling generation technique (which is described in the paper as well). evaluate.py is not the code used for the numbers reported in the paper, eval_accelerate.py is used. Please see the Readme for more details. I would be happy to clarify any other doubts as well. |
Thanks for your reply! But it seems that eval_accelerate.py just gives the unfiltered Hit@1, could you provide the official implementation for calculating MRR and filtered H@N? Appreciate your help! |
For that I would suggest you try the following:
I will be cleaning the code/adding documentation etc. for this soon. Meanwhile I suggest you try the scores + notebook approach. Let me know if this works. |
It seems to work great! Thank you so much for your patience! |
Hi, @apoorvumang, this work is so great!
I try to reproduce the reported MRR, MR and H@N results on Wikidata5M with evaluate.py, but I found it needs quite a long time. Could you please provide an estimation of the evaluation time for calculating MRR?
Thanks!
The text was updated successfully, but these errors were encountered: