-
Notifications
You must be signed in to change notification settings - Fork 89
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Tatoeba models outputting nonsense #35
Comments
This seems to be the case with all their models which originate from Tatoeba Challenge. Only the models which are included here seem to work using Hugging Face. Up until a month ago, I hadn't encountered such problems. |
Thanks for reporting, I'll try to check if the tokenizer or the model is wrong. |
Hey! you should use |
Are the opus-mt-xx-xx models a different issue? I tried just now on both old and newer transformers and haven't gotten them to work. |
Note that jap is not Japanese |
That makes more sense. |
On the huggingface demo, (e.g. https://huggingface.co/Helsinki-NLP/opus-tatoeba-en-ja?text=My+name+is+Wolfgang+and+I+live+in+Berlin) the output doesn't seem to make sense.
I ran some models locally too and this was the result of:
Output:
only opus-mt-ja-en gave an answer which was understandable at all. Any idea what the problem might be? The opus-mt-jap-en model also doesn't make a comprehensible translation.
The tatoeba models were converted to pytorch through
python -m transformers.models.marian.convert_marian_to_pytorch --src folder --dest folder-pytorch
I'm not sure how just pasting in the huggingface link loads it so I don't know how to replicate it.The text was updated successfully, but these errors were encountered: