-
Notifications
You must be signed in to change notification settings - Fork 241
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Any plans for QLora? #1537
Comments
Yes actually! We are currently working or rolling out some support for quantization generally and quantize lora. We will work on some actual docs for this soon, but for now you can check out the I believe the 4 bit normal float from the qlora paper is not currently available, but int8 quantized weights layered in with lora should be good to go. Early days on the feature, so any feedback appreciated as we build it out. |
This issue is stale because it has been open for 14 days with no activity. It will be closed if no further activity occurs. Thank you. |
Guide on this incoming... |
Hi @asmith26, Could you please confirm if this issue is resolved for you with the above comment ? Please feel free to close the issue if it is resolved ? Thank you. |
Have got this working https://keras.io/examples/keras_recipes/parameter_efficient_finetuning_of_gemma_with_lora_and_qlora/#qlora-finetuning Many thanks again! |
Hi there!
keras-nlp
supports Lora, e.g. from https://ai.google.dev/gemma/docs/lora_tuningJust wondering are there any plans to implement QLora finetuning, to help make finetuning even more accessible on smaller GPUs?
https://arxiv.org/abs/2305.14314
https://huggingface.co/blog/4bit-transformers-bitsandbytes
Many thanks for any help, and this lib!
The text was updated successfully, but these errors were encountered: