Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

api_demo模式部署后,在流式回复的前提下,能否支持多用户并发调用API? #3118

Closed
1 task done
RyanOvO opened this issue Apr 3, 2024 · 1 comment
Labels
solved This problem has been already solved.

Comments

@RyanOvO
Copy link

RyanOvO commented Apr 3, 2024

Reminder

  • I have read the README and searched the existing issues.

Reproduction

none

Expected behavior

  1. 目前发现api_demo部署后,调用API时,不支持多用户并发调用。(流式回复的场景)
  2. 如何实现多用户并发调用,模型可流式响应?

System Info

none

Others

none

@hiyouga
Copy link
Owner

hiyouga commented Apr 3, 2024

首页 readme 中 vllm 部署支持并发流式

@hiyouga hiyouga added the solved This problem has been already solved. label Apr 3, 2024
@hiyouga hiyouga closed this as completed Apr 3, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
solved This problem has been already solved.
Projects
None yet
Development

No branches or pull requests

2 participants