Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

llama.cpp GPU support #46

Open
alexl83 opened this issue May 14, 2023 · 1 comment
Open

llama.cpp GPU support #46

alexl83 opened this issue May 14, 2023 · 1 comment

Comments

@alexl83
Copy link

alexl83 commented May 14, 2023

Hi, since commit 905d87b70aa189623d500a28602d7a3a755a4769
llama.cpp support GPU inference with nvidia CUDA via command-line switches like --gpu-layers
Could you please consider adding support to GPT-LLAMA.CPP aswell?

Thank you!

@alexl83 alexl83 changed the title llama.cpp GPU support! llama.cpp GPU support May 14, 2023
@msj121
Copy link

msj121 commented Jun 11, 2023

@alexl83
Just looking at the code, since you compile llama.cpp in theory it would appear to me that you can install with cuda support, then you are just passing the argument like any of the others listed, like threads.

ie:
npm start ngl 4 to hoist on to 4 gpu layers. I don't have a compatible setup to test though.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants