Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

ChatGLM3的lora微调问题 #26

Closed
zhangmaohong opened this issue Jan 2, 2024 · 9 comments
Closed

ChatGLM3的lora微调问题 #26

zhangmaohong opened this issue Jan 2, 2024 · 9 comments

Comments

@zhangmaohong
Copy link

1、loss下降过快,但是没有定位到原因,训练结束也没有生成新的模型文件
loss异常
2、前面都能泡通,模型推理会抱错,但估计也是模型文件没有真正生成的原因导致的

@Hongru0306
Copy link
Contributor

请问用的是repo里面的数据吗?还是自己的数据?

@zhangmaohong
Copy link
Author

用的repo嬛嬛那个数据集

@Hongru0306
Copy link
Contributor

您好,我这边刚拉下来跑完,没有出现问题:
image
LoRa微调本身就是不保存权重的,它只保存lora微调的部分,加载的时候需要peft进行二者一起加载,细节可参考同目录下的md文件:
image

@zhangmaohong
Copy link
Author

我之前都是在notebook里面跑的 现在跑的py文件还是一样呢
image

@KMnO4-zx
Copy link
Contributor

KMnO4-zx commented Jan 2, 2024

你应该是前面的某一步搞错了,我们这边复现的结果loss是逐步下降的。请检查你之前的步骤。

@KMnO4-zx KMnO4-zx closed this as completed Jan 3, 2024
@sjy
Copy link

sjy commented Jan 4, 2024

我也遇到了一样的问题,按照文档跑的,loss 没有下降

@rxy1212
Copy link

rxy1212 commented Jan 15, 2024

我出现了跟楼主一样的问题,也是loss变成了0.0,也没有生成模型文件
image

@rxy1212
Copy link

rxy1212 commented Jan 15, 2024

将 peft 降级至 0.6.2 可以解决问题

@zjk000
Copy link

zjk000 commented Feb 9, 2024

你好请问一下你训练完之后是如何保存lora的权重到本地的?

llm = AutoModelForCausalLM.from_pretrained("THUDM/chatglm3-6b", trust_remote_code=True).cuda()
model = get_peft_model(llm, lora_config).cuda()
model.save_pretrained("trained_lora_weights")

请问是使用类似上述的代码保存的吗?我这么写有问题吗?为什么无法保存lora权重到本地?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

6 participants