-
Notifications
You must be signed in to change notification settings - Fork 2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Usage] finetune_task_lora.sh checkpoints usage #1423
Comments
I also got the problem when I try to use merge_lora_weights.
Did you solve? |
Check this out |
Hi leechangdong, did you modify the code beyond #844 . Currently, I change the trainer as #844, but only have I don't have the gradient information like the folder of "global_step*" as you show. |
@leechangdong Thank you! But I have another problem.I follow #844 modify my code, but it not save files as expected.And when I examine what errors happened, I find any "print" or "rank0_print"would not successfule print,it seems that it does not execute my local train.py/LLaVATrainer.py |
Describe the issue
Issue:
�train args:
finetune_task_lora.sh
With that training setup, I fine tune using my custom data, and checkpoints were saved for each step.
The folder contents of those "checkpoints" are as shown in the photo above, and in order to use those checkpoints for inference, I think need to run merge_lora. "scripts/merge_lora_weights.py" to use the above checkpoints for inference. and i did.
command:
But I get an error like this
My purpose is to use the checkpoints I saved while training "finetune_task_lora.sh" for inference.
Does anyone know if what I did above is correct and how to fix this error?
The text was updated successfully, but these errors were encountered: