Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Question] Is GGUF model package format supported with quantized models? #2227

Closed
qdrddr opened this issue Apr 26, 2024 · 3 comments
Closed
Labels
question Question about the usage

Comments

@qdrddr
Copy link

qdrddr commented Apr 26, 2024

❓ General Questions

Hi, can I use existing gguf model package format of quantized models such as this one?
https://huggingface.co/MaziyarPanahi/Mixtral-8x22B-Instruct-v0.1-GGUF

If not, can I convert it to MLC?

@qdrddr qdrddr added the question Question about the usage label Apr 26, 2024
@qdrddr qdrddr changed the title [Question] Is GGUF quantized models supported? [Question] Is GGUF model package format supported with quantized models? Apr 26, 2024
@Hzfengsy
Copy link
Member

GGUF is not supported with MLC-LLM. You could convert weight from the original model. See https://llm.mlc.ai/docs/compilation/convert_weights.html

@qdrddr
Copy link
Author

qdrddr commented Apr 29, 2024

Sorry for the stupid question, but aren't Hugging Face and GGUF two different formats?
I am specifically interested in GGUF and the manual you provided for HF format. Or are both supported by the converter? @Hzfengsy

@Hzfengsy
Copy link
Member

Yes Huggingface and GGUF are different formats. GGUF is not supported by the converter now

@qdrddr qdrddr closed this as completed May 1, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Question about the usage
Projects
None yet
Development

No branches or pull requests

2 participants