New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
the evaluation on the PCBA dataset seems wrong #92
Comments
Thanks for using Graphormer. In v1: the average precision is correctly calculation by gathering all results from different cards. Please kindly use this for Graphormer-v1 on PCBA. In v2: currently we haven't prepared pcba script in example since the architecture has been modified and we plan to release the pcba example after we search optimal configuration and hyper-parameters. If this feature is urgent for you, please kindly click the thumb up reaction at this issue, and we will promote the priority. |
Hi zhengsx, thank you for your replying. I'm a little confused now, since in the function validation_epoch_end of model.py, the code goes like: |
Ok, I got it, thank you! |
Hi authors,
Thank you for your great work! I noticed that the result of 'the mean of the 4 cards ap' is different from the result of 'gather all pred and labels of different cards and do evaluation once'. And the latter method's result tends to be lower than the former one. It seems that when doing evaluation, Graphormer is using the former method. May I know that if you have the valid and test result of Graphormer model evulating the whole dataset once? Thank you!
The text was updated successfully, but these errors were encountered: