Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Examples for Llama model architecture #2

Closed
okpatil4u opened this issue Sep 14, 2023 · 5 comments
Closed

Examples for Llama model architecture #2

okpatil4u opened this issue Sep 14, 2023 · 5 comments

Comments

@okpatil4u
Copy link

okpatil4u commented Sep 14, 2023

Hello Eric, this looks like great work ! Thank you !!

Can you please add examples for both training and inference for Llama model using candle-lora ? Is it supported through this work ?

@EricLBuehler
Copy link
Owner

Yes! With my candle-lora-macro library, all you need to do is derive the AutoLoraConvert and add the replace_layer_fields to all model structs of a LLama model. They will replace the concrete types and automate the conversion process. Then, call the conversion method on each model struct.

I plan on adding an example shortly. If you have any questions, let me know!

@okpatil4u
Copy link
Author

okpatil4u commented Sep 14, 2023 via email

@EricLBuehler
Copy link
Owner

Yes, once you convert to a LoRA model you could fine-tune it. After finetuning, you can merge the weights to speed up inference.

@EricLBuehler
Copy link
Owner

I am closing this so that it does not become a stale issue, but feel free to reopen. I will be adding a LoRA example for Llama and soon!

@okpatil4u
Copy link
Author

okpatil4u commented Sep 15, 2023 via email

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants