From 7a0b1606bae4ca58c901a36a5b77f89bdc076155 Mon Sep 17 00:00:00 2001 From: Tyler Michael Smith Date: Thu, 25 Sep 2025 12:33:00 -0400 Subject: [PATCH] rm max_completion_tokens Signed-off-by: Tyler Michael Smith --- src/guidellm/backend/openai.py | 1 - 1 file changed, 1 deletion(-) diff --git a/src/guidellm/backend/openai.py b/src/guidellm/backend/openai.py index 680578cc..759665bb 100644 --- a/src/guidellm/backend/openai.py +++ b/src/guidellm/backend/openai.py @@ -468,7 +468,6 @@ def _completions_payload( max_output_tokens or self.max_output_tokens, ) payload["max_tokens"] = max_output_tokens or self.max_output_tokens - payload["max_completion_tokens"] = payload["max_tokens"] if max_output_tokens: # only set stop and ignore_eos if max_output_tokens set at request level