Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Donut model only works on 4.36.2 for inference, not 4.37.2 #28846

Closed
2 of 4 tasks
VikParuchuri opened this issue Feb 2, 2024 · 4 comments
Closed
2 of 4 tasks

Donut model only works on 4.36.2 for inference, not 4.37.2 #28846

VikParuchuri opened this issue Feb 2, 2024 · 4 comments

Comments

@VikParuchuri
Copy link

System Info

I have a donut model with some slight customizations to the mbart decoder (added gqa and moe). It works fine on 4.36.2 and 4.37.2 for training.

But inference only works on 4.36.2. When I run inference on 4.37.2, then the output degenerates into repetition. @amyeroberts

Here is an example (the text has been ocred with donut, then rendered back onto a page image):

This is with 4.36.2:

image

And this is with 4.37.2:

image

Everything else is identical (same system, same packages). I don't see anything obvious in the release notes that would cause this.

Who can help?

No response

Information

  • The official example scripts
  • My own modified scripts

Tasks

  • An officially supported task in the examples folder (such as GLUE/SQuAD, ...)
  • My own task or dataset (give details below)

Reproduction

You should see different output with different transformers versions.

Expected behavior

I expect the output to be the same with both versions, and to not degenerate into repetition.

@ArthurZucker
Copy link
Collaborator

fyi @NielsRogge

Copy link

github-actions bot commented Mar 4, 2024

This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread.

Please note that issues that do not follow the contributing guidelines are likely to be ignored.

@NielsRogge
Copy link
Contributor

Hi sorry for the late reply, I've checked all slow integration tests of Donut, and they all pass:

RUN_SLOW=yes pytest tests/models/vision_encoder_decoder/test_modeling_vision_encoder_decoder.py::DonutModelIntegrationTest

Which would mean that Donut works as expected on v4.38.2 dev.

This might be explained because of the customizations? Could you clarify which things are changed compared to Donut?

Copy link

github-actions bot commented Apr 2, 2024

This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread.

Please note that issues that do not follow the contributing guidelines are likely to be ignored.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants