-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
GPU usage #15
Comments
Although I had nothing to do with the creation of this project, I see that it is using llama.cpp, a project that focuses on using CPU vectorization to run the model so my best guess is that no GPU is needed. |
I see. Actually I have an RTX 4090, so I was hoping I can use it, to speed things up. |
Same here. I hope that if GPU use is not yet supported, it will be soon. |
@AntouanK @farrael004 you guys can take a look at this issue where they discuss inference in consumer-grade GPUs meta-llama/llama#4 |
Hi there.
Thanks for sharing this project.
Does this use the GPU? or just the CPU?
The text was updated successfully, but these errors were encountered: