-
-
Notifications
You must be signed in to change notification settings - Fork 10.5k
Add AWQ support for all models #1714
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
FYI, Phi is not yet supported in AutoAWQ as there are some blockades associated with their architecture that have yet to be resolved. GPT-2 is also not supported yet, although I can add it if the community wants it. |
Hi @casper-hansen, thanks for letting us know! Could you elaborate more on this?
I thought the Phi's architecture is not very different from other GPT models. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM! Thanks for the quick fix!
Fixes #1682
This PR adds AWQ support for all models, by adding
ScaledActivation
.I didn't test AWQ for GPT2, GPT-NeoX, and Phi models, since I couldn't find their quantized weights in HF model hub.