Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Run evaluation script with bert sklearn wrapper to check metrics are consistent with run_squad script #70

Closed
fmikaelian opened this issue Mar 8, 2019 · 2 comments

Comments

@fmikaelian
Copy link
Collaborator

fmikaelian commented Mar 8, 2019

@andrelmfarias How did you generate metrics.json previously?

@fmikaelian fmikaelian changed the title Run evaluation script bert sklearn wrapper to check metrics are consistent with run_squad script Run evaluation script with bert sklearn wrapper to check metrics are consistent with run_squad script Mar 8, 2019
@fmikaelian
Copy link
Collaborator Author

fmikaelian commented Mar 11, 2019

To be added in README:

python evaluate-v1.1.py data/dev-v1.1.json logs/bert_qa_squad_v1.1_sklearn/predictions.json

@andrelmfarias
Copy link
Collaborator

andrelmfarias commented Apr 16, 2019

Running predictions for the model version with sklearn wrapper on the squad development set showed metrics comparable to the official hugging face version:

{"exact_match": 81.2488174077578, "f1": 88.43242225358777}

This shows the sklearn wrapper is well implemented and works correctly.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment