Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Feature Request: Support for Meta Chameleon 7B and 34B #7995

Closed
4 tasks done
arch-btw opened this issue Jun 18, 2024 · 12 comments
Closed
4 tasks done

Feature Request: Support for Meta Chameleon 7B and 34B #7995

arch-btw opened this issue Jun 18, 2024 · 12 comments
Labels
enhancement New feature or request stale

Comments

@arch-btw
Copy link
Contributor

arch-btw commented Jun 18, 2024

Prerequisites

  • I am running the latest code. Mention the version if possible as well.
  • I carefully followed the README.md.
  • I searched using keywords relevant to my issue to make sure that I am creating a new issue that is not already open (or closed).
  • I reviewed the Discussions, and have a new and useful enhancement to share.

Feature Description

"Meta Chameleon is a family of models that can combine text and images as input and output any combination of text and images with a single unified architecture for both encoding and decoding. While most current late-fusion models use diffusion-based learning, Meta Chameleon uses tokenization for text and images. This enables a more unified approach and makes the model easier to design, maintain, and scale. The possibilities are endless—imagine generating creative captions for images or using a mix of text prompts and images to create an entirely new scene."

Motivation

This would be a great addition to llama.cpp!

The image features look interesting but it can also simply do Text -> Text and a lot of other combinations:

  • Text -> Text
  • Image -> Image
  • Text -> Image
  • Image -> Text
  • Image -> Text + Image
  • Text + Image -> Text
  • Text -> Text + Image
  • Text + Image -> Text + Image
chameleon.mp4
@arch-btw arch-btw added the enhancement New feature or request label Jun 18, 2024
@EliEron
Copy link

EliEron commented Jun 18, 2024

Here are some relevant links:

Given it's a completely new architecture, and a multimodal one at that, I imagine adding support for it will not be easy. But I'm also very excited to see this supported.

Edit: According to Meta researcher Armen Aghajanyan the architecture is actually similar:

Similar architecture to LLaMa (apart from QK-norm), get fast inference working.

@arch-btw arch-btw changed the title Feature Request: Support for Meta Chameleon 7B and 30B Feature Request: Support for Meta Chameleon 7B and 34B Jun 18, 2024
@0wwafa
Copy link

0wwafa commented Jun 18, 2024

Yes! Please do! Also because as of now there is no way to run it on CPU only.

@SolvAI
Copy link

SolvAI commented Jun 19, 2024

yum yum yum :p

@ann-brown
Copy link

Since it uses similar to Medusa architecture is that likely to be supportable at the same time for the self-speculative decoding side of inference? It sounded like it could run without that, but it'd be neat to have that available too.

@jacobkahn
Copy link

Let me know if we can answer any questions about the architecture, inference, etc. Our reference implementation in https://github.com/facebookresearch/chameleon should be clear. Differences from the Llama architecture are minor:

@typedrat
Copy link

Considering the VQGAN is public, it should be possible for llama.cpp to reinstate the image output capabilities.

@chigkim
Copy link

chigkim commented Jun 25, 2024

+1000! I'd love to run Chameleon with llama.cpp!

@nopperl nopperl mentioned this issue Jul 17, 2024
4 tasks
@freeqaz
Copy link

freeqaz commented Jul 18, 2024

In the interim: Is anybody aware of any quantized weights for this model currently? Would love to try to 30b version on my 2x 3090s.

Found this issue because I know llama.cpp is a common way to shrink these models down.

@github-actions github-actions bot added the stale label Aug 18, 2024
Copy link
Contributor

github-actions bot commented Sep 1, 2024

This issue was closed because it has been inactive for 14 days since being marked as stale.

@github-actions github-actions bot closed this as completed Sep 1, 2024
@chigkim
Copy link

chigkim commented Sep 1, 2024

Still not supported.

@arch-btw
Copy link
Contributor Author

Hi guys, it got merged: #8543

@chigkim
Copy link

chigkim commented Sep 28, 2024

Hi guys, it got merged: #8543

The pr is only for text, not image right?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request stale
Projects
None yet
Development

No branches or pull requests

9 participants