New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Poor triple extractor performance (OpenIE) #17
Comments
Hey. How did you go about providing your test file? Is it by adding the file path to the gold parameter in OIE_2016.json? It seems to require some parameters additional to a sentence. |
Hi! I thought you can try more sentences when you test it, cause fewer examples can lead to bias. Then, you can use the evaluation metrics the paper mentioned to test whether the result is poor! Thanks! |
So the real problem with this was that I was looking at the raw triples before the ranking algorithm was applied. However, looking at the ranked triples, more second-guessing is involved. This output is ranked by the contrastive distance and hence only top (or bottom) triples should be taken seriously. |
@filip-cermak But how do you do that? Do you need to run your own data by some sort of dictionary? |
I followed the README and successfully run the OpenIE16 benchmark, then I modified
OIE_2016.json
file to point to my directory withtest.txt
file containing just one lineJulia owns two cats and one dog.
The output is however really poor, the expected triples[Julia, owns, two cats]
and[Julia, owns, one dog]
have low scores and there are many other (ill-created) triples, sometimes with even higher score values (complete output attached below).Is this the normal behavior of the model? Why is the performance so poor? Is there a systematic issue with how I am doing this?
The text was updated successfully, but these errors were encountered: