-
Notifications
You must be signed in to change notification settings - Fork 9.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Feature Request: Support for Meta Chameleon 7B and 34B #7995
Comments
Here are some relevant links:
Edit: According to Meta researcher Armen Aghajanyan the architecture is actually similar:
|
Yes! Please do! Also because as of now there is no way to run it on CPU only. |
yum yum yum :p |
Since it uses similar to Medusa architecture is that likely to be supportable at the same time for the self-speculative decoding side of inference? It sounded like it could run without that, but it'd be neat to have that available too. |
Let me know if we can answer any questions about the architecture, inference, etc. Our reference implementation in https://github.com/facebookresearch/chameleon should be clear. Differences from the Llama architecture are minor:
|
Considering the VQGAN is public, it should be possible for |
+1000! I'd love to run Chameleon with llama.cpp! |
In the interim: Is anybody aware of any quantized weights for this model currently? Would love to try to 30b version on my 2x 3090s. Found this issue because I know llama.cpp is a common way to shrink these models down. |
This issue was closed because it has been inactive for 14 days since being marked as stale. |
Still not supported. |
Hi guys, it got merged: #8543 |
The pr is only for text, not image right? |
Prerequisites
Feature Description
Motivation
This would be a great addition to llama.cpp!
The image features look interesting but it can also simply do Text -> Text and a lot of other combinations:
chameleon.mp4
The text was updated successfully, but these errors were encountered: