-
Notifications
You must be signed in to change notification settings - Fork 162
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
When running mii.serv, it keeps print waiting for server to start. #361
Comments
@cninnovationai could you please try loading the model with import mii
pipe = mii.pipeline("/data/python_workspace/text-generation-webui/models/neural-chat-7b-v3-1/")
response = pipe("test") |
[2023-12-26 14:40:25,529] [INFO] [real_accelerator.py:161:get_accelerator] Setting ds_accelerator to cuda (auto detect) always stay here,No further progress, gpu memory usage is 0 Do you know how to solve this problem? @mrwyattii |
@cninnovationai it looks like there is a lock file in your torch cache that is causing the issue. Try deleting your torch cache and run again: |
Cleared torch cache and worked for me |
after rm -rf /home/bruce/.cache/torch*,it work |
My OS is ubuntu 22.04
(base) bruce@bruce:~$ nvcc --version nvcc: NVIDIA (R) Cuda compiler driver Copyright (c) 2005-2022 NVIDIA Corporation Built on Wed_Sep_21_10:33:58_PDT_2022 Cuda compilation tools, release 11.8, V11.8.89 Build cuda_11.8.r11.8/compiler.31833905_0
run the code above, it keeps print waiting for server to start, but there is no any ERROR, and when run watch -n 1 nvidia-smi , GPU Memory Usage is zero.
The server has been unable to complete startup,Does anyone know what this problem is? How to solve it?
The text was updated successfully, but these errors were encountered: