Skip to content

Issues: vllm-project/vllm

[Roadmap] vLLM Roadmap Q1 2025
#11862 opened Jan 8, 2025 by simon-mo
Open 7
[V1] Feedback Thread
#12568 opened Jan 30, 2025 by simon-mo
Open 63
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Assignee
Filter by who’s assigned
Sort

Issues list

[Feature]: 使用guided_decoding来实现Function Calling feature request New feature or request
#14890 opened Mar 16, 2025 by shell-nlp
1 task done
[Usage]: ModuleNotFoundError: No module named 'triton' usage How to use vllm
#14888 opened Mar 16, 2025 by Hinsael
1 task done
[Bug]:0.74 dev ,the error occurred in the gptq_marlin_gemm function call bug Something isn't working
#14887 opened Mar 16, 2025 by su400
1 task done
[Bug]: RuntimeError: CUDA error: invalid argument bug Something isn't working
#14885 opened Mar 16, 2025 by vladlen32230
1 task done
[New Model]: Command A with tool support new model Requests to new models
#14866 opened Mar 15, 2025 by Hexoplon
1 task done
[Bug]: TTFT Performance Regression in vLLM v0.7.0 Compared to v0.6.1.post2 bug Something isn't working
#14845 opened Mar 14, 2025 by asleepykitty
1 task done
[Feature]: specify model only in config.yaml feature request New feature or request good first issue Good for newcomers
#14819 opened Mar 14, 2025 by g0t4
1 task done
[Usage]: max_model_len, max_num_seqs and mm_counts usage How to use vllm
#14816 opened Mar 14, 2025 by TheFloHub
[Bug]: Gemma-3-27b-it-GPTQ Can't run in sm75, vllm-0.7.4.dev bug Something isn't working
#14814 opened Mar 14, 2025 by HelloCard
1 task done
[Usage]: how to reduce the number of processes of compile_worker usage How to use vllm
#14808 opened Mar 14, 2025 by FanYaning
1 task done
[Bug]: CUDA_VISIBLE_DEVICES is not supported bug Something isn't working
#14807 opened Mar 14, 2025 by chenhongyu2048
1 task done
[Usage]: How to make sure the timeout takes effect usage How to use vllm
#14792 opened Mar 14, 2025 by DayDayupupupup
1 task done
[V1][Bug] IMA with ngram spec decoding and flashinfer bug Something isn't working
#14765 opened Mar 13, 2025 by markmc
1 task done
ProTip! Adding no:label will show everything without a label.