Skip to content

Issues: vllm-project/vllm

[Roadmap] vLLM Roadmap Q2 2024
#3861 opened Apr 4, 2024 by simon-mo
Open 29
v0.4.3 Release Tracker
#4895 opened May 18, 2024 by simon-mo
Open 13
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Label
Filter by label
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Milestones
Filter by milestone
Assignee
Filter by who’s assigned
Sort

Issues list

[Bug]: Model Launch Hangs with 16+ Ranks in vLLM bug Something isn't working
#5170 opened May 31, 2024 by wushidonguc
[Performance]: What can we learn from OctoAI performance Performance-related issues
#5167 opened May 31, 2024 by hmellor
[Bug]: Unable to Use KV Cache in AsyncLLMEngine bug Something isn't working
#5162 opened May 31, 2024 by kezouke
[Usage]: extractive question answering using VLLM usage How to use vllm
#5126 opened May 30, 2024 by suryavan11
[New Model]: LLaVA-NeXT-Video support new model Requests to new models
#5124 opened May 30, 2024 by AmazDeng
[Bug]: The tail problem bug Something isn't working
#5123 opened May 30, 2024 by ZixinxinWang
[Bug]: Can't run vllm distributed inference with vLLM + Ray bug Something isn't working
#5094 opened May 29, 2024 by linchen111
[Bug]: Gemma model fails with GPTQ marlin bug Something isn't working
#5088 opened May 28, 2024 by arunpatala
ProTip! no:milestone will show everything without a milestone.