Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Pre-trained Model #66

Closed
travellertea opened this issue Jun 2, 2018 · 1 comment
Closed

Pre-trained Model #66

travellertea opened this issue Jun 2, 2018 · 1 comment
Labels

Comments

@travellertea
Copy link

Hello There,

Question 1:
url = 'https://storage.googleapis.com/chakki/datasets/public/models.zip'
What is this model trained on for NER tagging?

Question 2:
Why is the trained model performing significantly better with the trained model, compared to when I train a model with Conll2003 data and evaluate on the test set?

Cheers,
TZ

@Hironsan
Copy link
Owner

Hironsan commented Jun 3, 2018

Thank you for your question.
My answers are as follows:

Answer 1:

This model is based on Bidirectional LSTM-CRF.

This is trained by CoNLL2003 dataset.

Answer 2:

Because the published model is trained by CoNLLL2003 all(train + valid + test) datasets.

In anaGo 1.0.0, I released new trained dataset. This is trained by CoNLL2003 train + valid datasets. I think it is comparable to some papers score.

with best regards

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

2 participants