Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[FEATURE REQUEST] GGUF inputs #4006

Closed
rcontesti opened this issue Jun 18, 2024 · 2 comments
Closed

[FEATURE REQUEST] GGUF inputs #4006

rcontesti opened this issue Jun 18, 2024 · 2 comments
Labels
feature A request for a proper, new feature. triaged This issue has been looked at a team member, and triaged and prioritized into an appropriate module

Comments

@rcontesti
Copy link

Llama.cpp has proven quite convininent for many of us, I wish you could add the possibility of taking those models as inputs. Many thanks!

@cccclai cccclai added the feature A request for a proper, new feature. label Jun 18, 2024
@mergennachin
Copy link
Contributor

@rcontesti - we have a work-in-progress prototype to convert fp16/fp32 version of llama

https://github.com/pytorch/executorch/tree/main/extension/gguf_util

@guangy10 guangy10 added the triaged This issue has been looked at a team member, and triaged and prioritized into an appropriate module label Jun 18, 2024
@rcontesti
Copy link
Author

Once again many thanks @mergennachin

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
feature A request for a proper, new feature. triaged This issue has been looked at a team member, and triaged and prioritized into an appropriate module
Projects
None yet
Development

No branches or pull requests

4 participants