-
Notifications
You must be signed in to change notification settings - Fork 780
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support the Mistral AI official huggingface weights for Mixtral-8x7B-v0.1 #81
Comments
Yea it's a good point. Will look into updating the example to use the official weights instead of the ones from MixtralKit |
One major issue how to run it with 64 GB memory ? This is a huge model, we need to think of reducing the precision. Is it also possible to think of delta in the weights between the chat (instruction version Mixtral-8x7B-Instruct-v0.1) and the native generator (mixtral-8x7b-32kseqlen). But I don't see an easy way so far to run on M1 64 GB. running the convert give me a terminal killed message
|
I was able to use the llama.cpp convert script to get it to seeing very weird output from the model though. will try to look into this more later, looking forward to learning/playing around w/ mlx on some of these models |
I'm able to get the someone13574 running. I attempted the cat/convert process on the official instruct model, but no dice. Is there a way to modify the convert script to get the instruct version going? |
This one right? https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1 I'll test it after getting the official Mistral one working |
That's the one. |
per the current example:
I'd really rather not have to re-download many GBs of weight files, and less-so from
someone13574
(no offense) when there weights are posted by Mistral AI itself: https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1would this work already? could the example be updated to use these weights?
The text was updated successfully, but these errors were encountered: