New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Question] Cross-lingual Score #7
Comments
Hello, we also conjecture that this is possible although we have not done a proper study regarding this hypothesis. |
Hi @Tiiiger, I was trying this:
And the confusion matrix plot is this: Clearly this is not working as the last words |
We just saw this https://github.com/facebookresearch/XLM implementation of a cross-lingual Language Model based on BERT.
There is an example on how it works. Could you consider implementing this in your library? |
Thank you @shoegazerstella for letting us know. We are definitely going to look into it but it may take some time before we get back to you. If this is really important to your research, I encourage you to fork the repo and start implementing it. The general backend of BERTScore is at https://github.com/Tiiiger/bert_score/blob/master/bert_score/utils.py. Please let me know if you have any questions. |
I am trying to implement the solution discussed. You can find the code here. Apologies if this is not the most elegant solution but was the fastest for me to test today. So I am running
I am printing the size of the tensors to compare with your implementation. Is this error related to their shape?
so here instead, |
I think you gave the wrong number of arguments to I will add documentation to |
Hope the added docs can help you. |
Hi @Tiiiger, thanks a lot for the docs, it helped a lot. As of now I have these very bad results:
Seems that |
Just a heads up. The absolute score may be less meaningful because they can have different ranges. Ideally, you would like to show that the Score correlates with human judgment, which unfortunately I don't know any. As I understand, you don't have any more implementation questions so I am closing this. I am happy to chat about the potential research opportunities for the cross-lingual scores. Feel free to continue the conversation under this issue or contact us directly through emails if you want to keep it private. |
Hi @Tiiiger, I had to modify something in the XLM vocabulary and use the bpe as tokenizer. You can see some more changes here. The code still does not work for more than 1 reference phrase at a time though. I would like to ask you a couple of questions on some things I am missing:
Thanks a lot for your help! |
Assumed that the embeddings have learned joint languages representations (so that
cat
is closer tokatze
orchat
, hence a sentence likeI love eating
will be closer toIch esse gerne
, as it happens in the MUSE or LASER models), would it be possible to evaluate the BERTscore against sentences in two different languages?The text was updated successfully, but these errors were encountered: