Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

ollama can't run qwen:72b, error msg ""gpu VRAM usage didn't recover within timeout #4427

Closed
changingshow opened this issue May 14, 2024 · 4 comments
Assignees
Labels
bug Something isn't working

Comments

@changingshow
Copy link

What is the issue?

I have already downloaded qwen:7b, but when i run ollama run qwen:7b,got this error Error: timed out waiting for llama runner to start:, in the server.log have this msg gpu VRAM usage didn't recover within timeout

OS

Windows

GPU

Nvidia

CPU

Intel

Ollama version

ollama version is 0.1.37

@changingshow changingshow added the bug Something isn't working label May 14, 2024
@chrisoutwright
Copy link

chrisoutwright commented May 19, 2024

I also get it (with smaller Params) when running RTX 2080 TI OR GTX 1060 with codeqwen:chat and codegamma:instruct for Win10.
It stays in RAM but will have to copy to GPU RAM everytime after one chat POST.

The GPU RAM is not exceeding on them, so not sure why it timesout every time.

@dhiltgen dhiltgen self-assigned this May 21, 2024
@dhiltgen
Copy link
Collaborator

This was fixed in 0.1.38 via pr #4430. If you're still seeing problems after upgrading, please share your server log and I'll reopen.

@pamanseau
Copy link

pamanseau commented May 24, 2024

@dhiltgen I have the same issue with 0.1.38 with Linux
ollama.log

@dhiltgen
Copy link
Collaborator

@pamanseau from the logs you shared, it looks like the client gave up before the model finished loading, and since the client request was canceled, we canceled the loading of the model. Are you using our CLI, or are you calling the API? If you're calling the API, what timeout are you setting in your client?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

4 participants