Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Question about the "word coreference" part #2

Closed
yhcc opened this issue Nov 7, 2021 · 2 comments
Closed

Question about the "word coreference" part #2

yhcc opened this issue Nov 7, 2021 · 2 comments

Comments

@yhcc
Copy link

yhcc commented Nov 7, 2021

First of all, thank you to release the code. Really nice work, so glad to see advances without enumerating span combinations. But I have one question about the "word coreference" part. Since the one word may be assigned to different coference clusters, such as for "a b c d e", "a b c" and "c" are both coferences and belong to different cluseters (if 'c' is the head word for "a b c"), how will this "word coreference" stage deal with it?

@vdobrovolskii
Copy link
Owner

Hi, thanks for the interest!

I am not sure I completely understand the question. I'll try to answer the way I understand it, please clarify if needed.

First of all, coreference resolution with word representations works the same way as with span representations, i.e. each mention can be only assigned to one cluster.

Then, the assumption is that there is one-to-one correspondence between all the valid spans in the text and their head words. That is, in "A big black cat sat on a mat" there will three valid spans for Ontonotes: "a big black cat" -> "cat", "sat" -> "sat" and "a mat" -> "mat".

So when the coreference links are found between individual words, each word is assigned to only one coreferent cluster and is then converted to only one span, so there are unique non-overlapping coreferent clusters of spans.

Now there are (very few) cases when two spans in the original Ontonotes dataset share the same head word. Almost all of them are related to conjunction. In those cases "A" and "A & B" are different spans with the same head word, "A". In our implementation such cases were simply discarded from the training set, because they were few and we were able to perform well, even though we couldn't predict any of such cases during inference.

To adopt this system under requirements that such cases should work, I would try assigning an artificial "head word" to such spans, for instance, the conjunction itself.

@yhcc
Copy link
Author

yhcc commented Nov 13, 2021

Thanks for your answer. Your understanding is right, you have solved my question.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants