-
Notifications
You must be signed in to change notification settings - Fork 25.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Marian models broken with latest transformers #26271
Comments
Hey, this is related to Helsinki-NLP/Tatoeba-Challenge#35 and a duplicate of #26216. The conversion script was a bit faulty and models need a re-upload. I can try to open PRs for relevant models |
That would be greatly appreciated, thank you! Right now I am running on transformers 4.28 but I am sure many others are interested in using the 1k+ translation models on latest version. |
I pushed a lot of PR on the hub, if you are still seeing this feel free to ping me. I think some of the big models were not updated yet, we'll see |
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread. Please note that issues that do not follow the contributing guidelines are likely to be ignored. |
Hello, we have Marian-converted models that are experiencing this issue (1 and 2). Could anyone recommend a solution please? These models are known to work with transformers versions greater than 4.26.1 up to 4.30.2, after which garbage is returned by both using the model's generate() method and the transformers.pipeline. Many thanks! Edit: Sorry, just found #26216 which seems to be the correct place for this! |
no worries, I think @LysandreJik is planning to finish updating the models that were missed in the great refactor! |
System Info
Transformers >= 4.31(and maybe also older), including master.
Who can help?
@Joaoga I noticed that you merged an update to Marian models.
Information
Tasks
examples
folder (such as GLUE/SQuAD, ...)Reproduction
Thank you for your great work :)
I can confirm that pretrained Marian models are broken on transformers 4.31 and on master, and that they are working on 4.28.
For reproduction see for example the following snippet, taken from Helsinki-NLP/opus-mt-tc-big-he-en model card - on 4.28 I get the expected output, on 4.31 and master I get random gibberish. I don't know if this also happens with other models and languages but I suspect so.
Expected behavior
See above
The text was updated successfully, but these errors were encountered: