-
Notifications
You must be signed in to change notification settings - Fork 1.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG] <title> 'ZeRO3 is incompatible with LoRA when finetuning on base model.' #1104
Comments
Thanks for your help! I have read this README. But my question mainly focuses on why "if we have these parameters trainable, it is not available to use ZeRO 3". Thanks very much |
The Please note that we have previously emphasized that Qwen(1.0) codebase and models are no longer subject to further updates. Therefore, for access to the latest features and ongoing support, we advise users to migrate their work to Qwen1.5. |
✅ Got you, thanks for your kindly reply |
您好,我对Qwen-14b-chat模型进行lora微调,使用zero3,还是会出现这个问题。另外我在8张4090显卡上进行lora微调,设置zero2的时候,显存会溢出,这个多卡的时候,微调显存是怎样计算的?期待您的回答 |
是否已有关于该错误的issue或讨论? | Is there an existing issue / discussion for this?
该问题是否在FAQ中有解答? | Is there an existing answer for this in FAQ?
当前行为 | Current Behavior
'ZeRO3 is incompatible with LoRA when finetuning on base model.'
期望行为 | Expected Behavior
'ZeRO3 is incompatible with LoRA when finetuning on base model.'
复现方法 | Steps To Reproduce
No response
运行环境 | Environment
备注 | Anything else?
我想知道为什么代码当中需要有这一段 ,求解答
The text was updated successfully, but these errors were encountered: