Skip to content

Conversation

@RyanMullins
Copy link
Contributor

What does this PR do?

Maintenance on the Gemma 3 weights conversion script.

Before submitting

  • This PR fixes a typo or improves the docs (you can dismiss the other checks if that's the case).
  • Did you read the contributor guideline,
    Pull Request section?
  • Was this discussed/approved via a Github issue or the forum? Please add a link
    to it if that's the case.
    • Discussed on Slack.
  • Did you make sure to update the documentation with your changes? Here are the
    documentation guidelines, and
    here are tips on formatting docstrings.
  • Did you write any new necessary tests?

Who can review?

Anyone in the community is free to review the PR once the tests have passed. Feel free to tag
members/contributors who may be interested in your PR.

cc @ArthurZucker @Cyrilvallez @zucchini-nlp

@RyanMullins
Copy link
Contributor Author

Please hold off on merging this. Going to add one more flag.

@RyanMullins
Copy link
Contributor Author

Okay, flag added. Ready for review and merge at your leisure. Thanks for the patience 🤗

@github-actions
Copy link
Contributor

[For maintainers] Suggested jobs to run (before merge)

run-slow: gemma3

@HuggingFaceDocBuilderDev

The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.

Copy link
Member

@zucchini-nlp zucchini-nlp left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks , lgtm! Just curious if the new vocab size is an intended change?

@RyanMullins
Copy link
Contributor Author

Yes it is. Since we've had a few recent releases with text only variants (270M, EmbeddingGemma), we're normalizing on the original vocab size (262144) in the main configs and then adding 64 to that during conversion if the include vision encoder flag is true.

@zucchini-nlp
Copy link
Member

Oke, thanks for clarifying

@zucchini-nlp zucchini-nlp merged commit 8fc5420 into huggingface:main Oct 20, 2025
13 checks passed
ngazagna-qc pushed a commit to ngazagna-qc/transformers that referenced this pull request Oct 23, 2025
* conversion: add include_vision_encoder flag (default true)

* conversion: update for inverted model.language_model weight path

* conversion: revert include_vision_encoder to True by default

* conversion: add chat template path flag
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants