-
Notifications
You must be signed in to change notification settings - Fork 12.1k
dual GPU 8G/16G - CUDA error: out of memory with dolphin-mixtral #3460
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
I encounter the same issue from time to time when num_ctx is set to 2048. |
Same as this, someone is working on it..... |
Previously it was running well but after some time, started to show same error that :
then i checked "C:\Users<username>\AppData\Local\Ollama\server.log" file and found following error at end of file:
then i tried with following solution : but after this it is consuming too much RAM about 90% of my RAM!, but yah its running 👍 |
I don't have a test environment to verify this asymmetry, but PR #4517 may fix this. |
On my side it works fine now, I haven't being seeing this error how about a few weeks or maybe a month. |
I just checked and it "seems" to work with WebUI 0.2.2 and ollama 0.1.41 |
What is the issue?
Ollama crashes out entirely with error (throws error, then terminates process)
[CUDA error: out of memory current device: 0, in function alloc at C:\a\ollama\ollama\llm\llama.cpp\ggml-cuda.cu:445 cudaMalloc((void **) &ptr, look_ahead_size) GGML_ASSERT: C:\a\ollama\ollama\llm\llama.cpp\ggml-cuda.cu:193: !"CUDA error"](error: out of memory current device: 0, in function alloc at C:\a\ollama\ollama\llm\llama.cpp\ggml-cuda.cu:445 cudaMalloc((void **) &ptr, look_ahead_size) GGML_ASSERT: C:\a\ollama\ollama\llm\llama.cpp\ggml-cuda.cu:193: !"CUDA error")
What did you expect to see?
Output (any)
Steps to reproduce
Notes:
Are there any recent changes that introduced the issue?
Update from 0.1.29 to 0.1.30 (reverting back to 0.1.29 fixed)
OS
Windows
Architecture
amd64
Platform
No response
Ollama version
0.1.30
GPU
Nvidia
GPU info
GPU 0: NVIDIA GeForce GTX 1080 (8GB)
GPU 1: Tesla P100-PCIE-16GB
CPU
AMD
Other software
Windows Server 2022 Standard x64 21H2
The text was updated successfully, but these errors were encountered: