You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Oct 31, 2023. It is now read-only.
I was wondering whether the English multilingual word embeddings are just a normalized subset (200K) of the fastText English embeddings?
I can see that the vectors are different between the two, would be happy to know if the reason is normalization or something else :)
Thanks!
The text was updated successfully, but these errors were encountered:
Yes, the English embeddings will be the same as the normalized fastText embeddings. We realized later that the normalization can be a bit detrimental so we will probably remove it soon. If you comment out these 2 lines: https://github.com/facebookresearch/MUSE/blob/master/src/trainer.py#L249-L250
then the source embeddings should not be modified.
I was wondering whether the English multilingual word embeddings are just a normalized subset (200K) of the fastText English embeddings?
I can see that the vectors are different between the two, would be happy to know if the reason is normalization or something else :)
Thanks!
The text was updated successfully, but these errors were encountered: