-
Notifications
You must be signed in to change notification settings - Fork 58
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support fine-tuning LLaMA3? #264
Comments
@timturing
LLaMA3-8b has been fine-tuned according to the solution in this issue (Otherwise, the fine-tuned model cannot be stopped normally), but there are still some strange prefixes when inference, which may still be related to tokenizer. After all, according to the method in this issue, That is, tokenizer.add_eos_token = True or tokenizer.add_eos_token = False doesn't work. How should this be solved? Thank you. |
Thank you for providing the additional information. I have observed the issue as well. As you mentioned, this cannot be resolved with our current code. Should the transformers team develop a solution, we will promptly update our code to incorporate it. |
Ok. Closing this issue. Looking forward to fix. |
Very great project!
I tried to use
training/bash/run_ds3.sh
fine-tuning LLaMA3-8b. However, I found that the following code does not work for LLaMA-3 during the debugging process. Is this in line with expectations?LLMBox/training/dataset/sft_dataset/sftdataset.py
Line 64 in 1f4fe29
Thanks for your reply.
The text was updated successfully, but these errors were encountered: