-
Notifications
You must be signed in to change notification settings - Fork 27
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Is Alpaca 13B and 30B tested? #12
Comments
I will try and get it integrated tonight ;) |
Hi, You can look at this branch You will have to follow the build steps and convert the model again. The issue with LoRA models are their embedding size. Based on how LoRA method works (It creates low rank decomposition matrices and freezes the pertained weights), I suspect that is why we have have different embedding sizes compared to non LoRA models. Will need to sort out a few things before merging to main but feel free to use this and let me know if you face any issues :) |
Merged to main. Structure of |
Hi @PotatoSpudowski . I was curious how alpaca models are handled differently. For example, |
Yup, That's why why require users to specify the ModelIdentifier when initialising the model. The ins flag if I am not right is supposed to specify that it is in instruction mode is it? Either ways we have example files for Alpaca and LLaMA models which show how to use these models for either text completion or QNA tasks. Finally we also are working on redesigning our save and load feature and optimising it for latency and size in the Developers should be allowed to implement their own workflows using the features that were developed using first principles thinking rather than us deciding workflows for them. Will document everything extensively so it is easier for everyone!!! |
I tried to run setting the path from huggingface, didn't work for 13B and 30B version but worked for 7B version.
The text was updated successfully, but these errors were encountered: