-
Notifications
You must be signed in to change notification settings - Fork 125
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Moe finetune error #17
Comments
Could you post your command & transformers version & deepspeed version & torch version. Thanks |
Thanks for your reply. The script is MoE-LLaVA/scripts/v1_5/qwen/finetune_moe.sh torch version: 2.0.1+cu117 command: |
command
|
Have you try it on |
this error still occurred on |
Copy that. I will check it as soon as possible. |
Hi, everyone. Sorry for that, we updated the new runing command to fix it. Checking out here |
Works for me, thx~! |
For training stage 3, we print the loss for each step to observe the router banlanced loss. If someone do not like it, just comment out this line. |
I still have the following error
here is my command
here is my package version
|
I still have the following error !!!!! AssertionError: The model has moe layers, but None of the param groups are marked as MoE. Create a param group with 'moe' key set to True before creating optimizer here is my package version: |
hi, I have the same environment setting as you. And meet the same problem when training a moe version of llama_llava. Have you solve the problem now? |
The following error occurred while running the script finetune_moe.sh:
The model has moe layers, but None of the param groups are marked as MoE. Create a param group with 'moe' key set to True before creating optimizer.
The text was updated successfully, but these errors were encountered: