Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Low POS in WSJ #6

Closed
ava-YangL opened this issue Jul 16, 2019 · 3 comments
Closed

Low POS in WSJ #6

ava-YangL opened this issue Jul 16, 2019 · 3 comments

Comments

@ava-YangL
Copy link

Hi , I tested on the WSJ dataset with model256 and only got accuracy about 95.5%. I would like to ask that how can i get the accuracy 97.97 of the paper.
I used the parameters set in the code, no changes were made.

@datquocnguyen
Copy link
Owner

datquocnguyen commented Jul 16, 2019

You are using Python 2.7 & DyNet 2.0.3, aren't you ?
(Also, the Stanford conversion toolkit v3.3.0 to obtain the data ? )
What is the parsing accuracy you get ?

@ava-YangL
Copy link
Author

Thank you for your reply.
The version of Python and DyNet is right. But i didn't use Stanford conversion toolkit.
I just want to get the result of POS tagging, do I also need to use the Stanford Toolkit in this case?

@datquocnguyen
Copy link
Owner

No, you do not really need using the Stanford conversion toolkit. Note that for model256, the POS tags used are PTB POS tags, not UPOS tags.

But, using data produced by the Stanford conversion toolkit would help you evaluating the predicted output properly (and easier). I am pretty confident that you can reproduce the reported scores in this manner (I already tested/rerun few times the trained model before releasing it).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants