-
Notifications
You must be signed in to change notification settings - Fork 5.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Cannot Use GPU properly #4395
Comments
I find this relevant to the topic, so I hope it's ok to reuse this ticket: I am trying ollama (preview for Windows) on my Windows box with llama3. This runs on a Dell 9510 with an Intel GPU (0) and an NVidea GPU (1). I have installed CUDA before installing ollama (not sure if this makes any difference) Is this normal? The GPU usage doesn't go above 0.1%. Shouldn't the GPU be used more than the CPU? |
@applepieiris what version of Linux are you using? I have been using an a100 on Ubuntu 22.04 and it's working correctly. If you can upgrade to the newest version of ollama you can try out the
If you're on the CPU, it will say @mr-j0nes I think your issue is probably different than this one, however, you can use the same |
In addition to what Patrick mentioned, the log you included above seems to be from 0.1.31. Please do upgrade to pick up fixes we've made around GPU discovery, and you might also want to check out the troubleshooting notes here https://github.com/ollama/ollama/blob/main/docs/troubleshooting.md#container-fails-to-run-on-nvidia-gpu which might cover your scenario. (cudart init 3 failures are one symptom covered by the troubleshooting) If things still aren't working as expected and it's not running on the GPU, please share your server log so I can see what caused us not to load on GPU. |
I am having similar issue, I have server with Ubuntu 22.02 desktop with Nvidia 3070 and 3080 GPU. |
@algocrypto the server log will help us understand why it's not able to discover the GPU. https://github.com/ollama/ollama/blob/main/docs/troubleshooting.md |
Thanks for sharing the link, it seems that I have the Nvidia GPU/Cuda driver issue, it is fixed now using Nvidia driver 535 and Cuda toolkit 11.5 |
@applepieiris are you still having troubles after upgrading and following the nvidia troubleshooting linked above? |
What is the issue?
I installed the Ollama in my linux server according to the official documents:
curl -fsSL https://ollama.com/install.sh | sh
Installation is ok and it returns:
`
But when I
ollama run llama2
, when the model file downloaded already. The GPU shows no running process:But when I checked the CPU usages:
From the above, we can see that the ollama is running on CPU!!
I check the logs of ollama it shows me:
Is there any solutions to this?
OS
Linux
GPU
Nvidia
CPU
Intel
Ollama version
0.1.37
The text was updated successfully, but these errors were encountered: