Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

FEAT: Support Qwen MoE model for huggingface and modelscope #1263

Merged
merged 3 commits into from Apr 9, 2024

Conversation

xiaodouzi666
Copy link
Contributor

No description provided.

@XprobeBot XprobeBot added this to the v0.10.1 milestone Apr 8, 2024
@ChengjieLi28 ChengjieLi28 merged commit c0dbe48 into xorbitsai:main Apr 9, 2024
9 of 12 checks passed
@XiaoCC
Copy link

XiaoCC commented Apr 11, 2024

The checkpoint you are trying to load has model type qwen2_moe but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date.

@LucisBaoshg
Copy link

https://huggingface.co/Qwen/Qwen1.5-MoE-A2.7B

Requirements
The code of Qwen1.5-MoE has been in the latest Hugging face transformers and we advise you to build from source with command pip install git+https://github.com/huggingface/transformers

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

7 participants