-
-
Notifications
You must be signed in to change notification settings - Fork 1.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Strange grad_norm spikes with rsLoRA on LLaMA-3 #577
Comments
RSLoRA scales the lora matrices by |
Do you mean 5e-6 ? I'm already using even smaller value 1e-5, so not sure how higher might help. |
Sadly this isn't an Unsloth issue, but llama-3 being more sensitive to rsLoRA :( Another way is to reduce |
It sounds like you shouldn't configure Alpha at all with rsLora. https://huggingface.co/blog/damjan-k/rslora
|
@BugReporterZ Ye so the suggested alpha for RSLoRA is the same as the rank |
I always see not expected grad_norm spikes when training LLaMA-3 models with Unsloth and rsLoRA:
It's OK without rsLoRA:
It was OK too when I've trained LLaMA-2 models with Unsloth and rsLoRA.
So now I'm in doubt should I just give up rsLoRA or is it possible to fix?
The text was updated successfully, but these errors were encountered: