Skip to content

Conversation

@ggerganov
Copy link
Member

While looking into #17260, found this error in the logic:

The slot_batched could end up being released (for example if the prompt does not fit into the context). The fix is to set slot_batched ptr only after we have actually queued any tokens for that slot.

@ggerganov ggerganov merged commit d396b43 into master Nov 14, 2025
66 checks passed
@ggerganov ggerganov deleted the gg/server-fix-can-batch-with branch November 14, 2025 12:03
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants