-
Notifications
You must be signed in to change notification settings - Fork 28
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Mistral Support #81
Comments
Hi! Yes, Mistral 7B is on our radar but we don't have an implementation for it yet. Our |
In particular Mixtral (with x, the mixture of experts version) could benefit a lot from this. As Mistral and Mixtral are Apache licenced, you could share smaller sliced versions. |
Great suggestion, for MoEs we need to modify the method slightly to account for the different architecture - they won't work out of the box with our current adapters. The computational invariance on which SliceGPT relies still applies though, so they should be sliceable. |
I write mixtral implementation of slicegpt. Here is my forked repository, https://github.com/noah-kim-theori/TransformerCompression, |
Amazing, nice work @noah-kim-theori! Could you share some perplexity and zero-shot accuracies of a sliced mixtral at e.g. 25% slicing vs dense? |
Hi,
Thanks for releasing this work! Are there any plans to release a Mistral version?
Thanks!
The text was updated successfully, but these errors were encountered: