-
Notifications
You must be signed in to change notification settings - Fork 440
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
QUESTION: No available slot found for the model #888
Comments
It means GPU has been used by other launched models, there's no slot left. |
是的,我已经加载一个模型ChatGLM3-6B-32k,再次加载其他的模型时就会出现这个错误提示!!! |
意思是一块GPU只能运行一个模型?这有点浪费资源了吧 |
目前就是这样的…… |
持续关注,没想到会这样,先用回FastChat吧 |
我这边的复现逻辑
class StatusGuardActor(xo.StatelessActor):
def __init__(self):
super().__init__()
self._model_uid_to_info: Dict[str, InstanceInfo] = {} # type: ignore |
This issue is stale because it has been open for 7 days with no activity. |
This issue was closed because it has been inactive for 5 days since being marked as stale. |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Server error: 503 - [address=0.0.0.0:36042, pid=294316] No available slot found for the model
The text was updated successfully, but these errors were encountered: