-
Notifications
You must be signed in to change notification settings - Fork 2
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Predictions on MathQA using pre-trained model #4
Comments
Hello, you can try adjusting the learning rate to 5e-5, and then observe the results. |
@ChaoYan Thank you for the response. I am not trying to train or fine-tune the model, I am just trying to use the trained model (after stage 2 as mentioned), provided by the authors here, in order to get the predictions on the MathQA test-set. I was hoping the model I am using from the link is the final version of the one named "Monolingual Setting BERT-TD w CL" that achieves 76.3% accuracy in MathQA (with no need for further fine-tuning). If am using a wrong model or process for inference, I would be thankful If you could provide me with instructions on this. Or if there are available somewhere the predictions of the model on the test-set of the MathQA directly, I would be glad if you could share them. Once again, thank you for making your code public! |
For loading a model, you need to add " --model_reload_path model_monolingual_en " like |
Hello! Thanks for making your work and models available, alongside steps to reproduce the results. I follow your guideline, but I make mistakes. It's like the trained model is mismatched with the neural network structure. What should I do to run the code successfully? Looking forward to your reply and thanks! @zwx980624 python run_ft.py
Process finished with exit code 1 |
Hi @ICanFlyGFC ! I have received the same problem as you. could you solve it later? If you could share your experience with me, it would be so helpful, thanks :) |
Hello! Thanks for making available your work and models, alongside steps to reproduce the results.
I am trying to generate the predictions on the MathQA dataset, using the trained monolingual_en model that you provide us with in the google drive link.
Essentially, I am using the test_model found in run_ft with minor modifications for paths and for keeping the results from compute_prefix_tree_result.
However, the results I get are greatly lower than the ones report (accuracy at around 0.03%), so obviously I am not running something correctly.
What I am doing is similar to running something along the lines:
python run_ft.py
--output_dir model_monolingual_en
--bert_pretrain_path ./pretrained_models_vocab/bert-base-uncased
--data_dir data
--train_file MathQA_bert_token_train.json
--test_file MathQA_bert_token_test.json
--hidden_size 768
--only_test
(obviously I am loading/handling correctly the paths in the changed test_model for my purpose).
Is there something wrong with this approach? Should I be using a different model altogether?
Thank you for your time.
The text was updated successfully, but these errors were encountered: