Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Sharing Training Details #7

Closed
YusenZhang826 opened this issue Oct 14, 2021 · 3 comments
Closed

Sharing Training Details #7

YusenZhang826 opened this issue Oct 14, 2021 · 3 comments

Comments

@YusenZhang826
Copy link

Hi Pan,

Your paper "Inter-GPS: Interpretable Geometry Problem Solving with Formal Language and Symbolic Reasoning" is awesome.

However, when reproducing results of this work, I have one problem. I trained the symbol detection model with the data you provided, but the model could not perform as well as the box_result you released. Could you please share more training details?

Thank you very much and looking forward to your reply.

@lupantech
Copy link
Owner

Hi Yusen,

Thank you for following our work! The released code sets the default arguments that we used for training the symbol detection model. The symbol detection results are saved in box_results and not quantitatively evaluated in our experiment. So I was wondering how you evaluated your detection results and what they looked like?

Best,
Pan

@YusenZhang826
Copy link
Author

YusenZhang826 commented Oct 18, 2021

Hi Pan,

Thanks for your response. I trained the detection model following your default arguments, and evaluated the results on the test set using the function csv_eval.evaluate in your code. The evaluation results are shown in the figure evaluation.png, and the detection results on the test set are saved in box_results_reproduce.zip.box_results_reproduce.zip

And the accuracy of Symbolic Solver task is only around 26% when we replaced the origin box_results with box_results_reproduce(keeping other arguments fixed). Could you please let me know where the problem is.

Looking forward for your reply.

evaluation

@lupantech
Copy link
Owner

lupantech commented Oct 24, 2021

Hi Yusen,

Sorry for the late response! I trained the detection model and I obtained similar detection results as yours. So the detection model works well. 

I was wondering if you ran the text recognition tool MathpixOCR and generated the diagram logic forms again before you tested the symbolic solver. If you generate new symbol detection results, you must also repeat these two steps. MathpixOCR we used is not free and costs about $20 for our test dataset. So, if possible, you can replace MathpixOCR with other open-source tools or models.

Pan

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants