Skip to content

Conversation

WoosukKwon
Copy link
Collaborator

No description provided.

Signed-off-by: Woosuk Kwon <woosuk@thinkingmachines.ai>
@WoosukKwon WoosukKwon added the ready ONLY add when PR is ready to merge/full CI is needed label Sep 21, 2025
@mergify mergify bot added deepseek Related to DeepSeek models llama Related to Llama models qwen Related to Qwen models gpt-oss Related to GPT-OSS models speculative-decoding labels Sep 21, 2025
@mergify mergify bot added the v1 label Sep 21, 2025
@mergify mergify bot added the tpu Related to Google TPUs label Sep 21, 2025
Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request is a large-scale refactoring to remove the deprecated SamplingMetadata from compute_logits and other related functions. The changes are applied consistently across a large number of files, including model implementations, test files, and worker runners. The SamplingMetadata class and its imports are removed, simplifying the method signatures. The changes appear correct and are a good cleanup of the codebase.

@WoosukKwon WoosukKwon merged commit 1c3ffdb into main Sep 21, 2025
59 of 66 checks passed
@github-project-automation github-project-automation bot moved this from To Triage to Done in gpt-oss Issues & Enhancements Sep 21, 2025
@WoosukKwon WoosukKwon deleted the woosuk/rm-v0-sampl-metadata branch September 21, 2025 17:37
@hmellor hmellor moved this to Done in V0 Deprecation Sep 21, 2025
kingsmad pushed a commit to kingsmad/vllm that referenced this pull request Sep 22, 2025
Signed-off-by: Woosuk Kwon <woosuk@thinkingmachines.ai>
wangxiyuan pushed a commit to vllm-project/vllm-ascend that referenced this pull request Sep 22, 2025
### What this PR does / why we need it?
This pr bump vllm commit hash to
vllm-project/vllm@5aeb925
fix issues:  
1. vllm-project/vllm#25345 has remove v0
metadata
2. vllm-project/vllm#25332
3. vllm-project/vllm#25334
4. vllm-project/vllm#23558, note that this vllm
commit update the model register logic, which will check all the model
registered have the `vllm.model_executor.models` path , which breaks our
custom registration of the deepseek_v3 model (it doesn't exist in the
vllm model path). so I move deepseek_v3 model registy to deepseek_v2 to
solve temporary

### How was this patch tested?

- vLLM version: v0.10.2
- vLLM main:
vllm-project/vllm@9607d5e

---------

Signed-off-by: wangli <wangli858794774@gmail.com>
FeiDaLI pushed a commit to FeiDaLI/vllm that referenced this pull request Sep 25, 2025
Signed-off-by: Woosuk Kwon <woosuk@thinkingmachines.ai>
charlifu pushed a commit to ROCm/vllm that referenced this pull request Sep 25, 2025
Signed-off-by: Woosuk Kwon <woosuk@thinkingmachines.ai>
Signed-off-by: charlifu <charlifu@amd.com>
yewentao256 pushed a commit that referenced this pull request Oct 3, 2025
Signed-off-by: Woosuk Kwon <woosuk@thinkingmachines.ai>
Signed-off-by: yewentao256 <zhyanwentao@126.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
deepseek Related to DeepSeek models gpt-oss Related to GPT-OSS models llama Related to Llama models qwen Related to Qwen models ready ONLY add when PR is ready to merge/full CI is needed speculative-decoding tpu Related to Google TPUs v1
Projects
Status: Done
Development

Successfully merging this pull request may close these issues.

1 participant