Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

m1/2 gpu support #101

Closed
stoneLee81 opened this issue Jul 1, 2023 · 5 comments
Closed

m1/2 gpu support #101

stoneLee81 opened this issue Jul 1, 2023 · 5 comments

Comments

@stoneLee81
Copy link

Feature request

it seems that it just support Nvidia gpu and chatglm/chatglm2 can't run under apple silicon env

Motivation

No response

Other

No response

@alew3
Copy link

alew3 commented Jul 1, 2023

+1 for this!

CPU works, but GPU only models are currently looking for NVidia GPUs which are not supported on the mac.

@aarnphm
Copy link
Member

aarnphm commented Jul 1, 2023

I can enable chatglm on CPU, but there is really no point since the inference are just way to slow to run on CPU anw

@alew3
Copy link

alew3 commented Jul 2, 2023

@aarnphm Is there any chance of GPU support for the Mac on Metal? Tensorflow and Pytorch already support Metal (not sure of how extensively though) Since the Mac shares memory with the VRam, it would be much more viable running a LLM locally as the other alternative is buying a professional NVidia Card.

@aarnphm
Copy link
Member

aarnphm commented Jul 2, 2023

Support metal is irrelevant of this. Metal won't make model inference faster unless we implement the attention layer in bare metal code

@alew3
Copy link

alew3 commented Jul 3, 2023

thanks for the feedback, much appreciated!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants