Replies: 1 comment
-
You’ve identified a classic determinism failure—ProblemMap No.15: “Inference seed drift & non-reproducibility.”
This means that even with the same chat_template, you’ll hit small, often invisible, numerical differences that amplify down the decoding path. Problem details and mitigation tips are mapped in this public index: |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
We are using Qwen2.5-14B-Instruct with vLLM. However, we found the following things can make the output different, even we set
temperature=0,top_p=1,seed=42
:vllm serve
is different with vllm offline inference, using the same chat_templatevllm serve
with different number of cardsThat is strange. Can someone tell me why? and how can I fix the output, when changing inference enveriments?
Beta Was this translation helpful? Give feedback.
All reactions