-
Notifications
You must be signed in to change notification settings - Fork 582
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
m1/2 gpu support #101
Comments
+1 for this! CPU works, but GPU only models are currently looking for NVidia GPUs which are not supported on the mac. |
I can enable chatglm on CPU, but there is really no point since the inference are just way to slow to run on CPU anw |
@aarnphm Is there any chance of GPU support for the Mac on Metal? Tensorflow and Pytorch already support Metal (not sure of how extensively though) Since the Mac shares memory with the VRam, it would be much more viable running a LLM locally as the other alternative is buying a professional NVidia Card. |
Support metal is irrelevant of this. Metal won't make model inference faster unless we implement the attention layer in bare metal code |
thanks for the feedback, much appreciated! |
Feature request
it seems that it just support Nvidia gpu and chatglm/chatglm2 can't run under apple silicon env
Motivation
No response
Other
No response
The text was updated successfully, but these errors were encountered: