From b7fc7000431e948772df344d99f2a640fc38c369 Mon Sep 17 00:00:00 2001 From: kamilbenkirane Date: Thu, 11 Dec 2025 20:25:35 +0100 Subject: [PATCH] Add GPT-5.2, GPT-5.2 Pro, and GPT-5.2 Instant models --- .../providers/openai/models.py | 39 +++++++++++++++++++ 1 file changed, 39 insertions(+) diff --git a/packages/text-generation/src/celeste_text_generation/providers/openai/models.py b/packages/text-generation/src/celeste_text_generation/providers/openai/models.py index 76111a5..8dd2fb3 100644 --- a/packages/text-generation/src/celeste_text_generation/providers/openai/models.py +++ b/packages/text-generation/src/celeste_text_generation/providers/openai/models.py @@ -71,6 +71,45 @@ TextGenerationParameter.OUTPUT_SCHEMA: Schema(), }, ), + Model( + id="gpt-5.2", + provider=Provider.OPENAI, + display_name="GPT-5.2", + streaming=True, + parameter_constraints={ + Parameter.MAX_TOKENS: Range(min=1, max=128000), + TextGenerationParameter.THINKING_BUDGET: Choice( + options=["minimal", "low", "medium", "high", "xhigh"] + ), + TextGenerationParameter.OUTPUT_SCHEMA: Schema(), + }, + ), + Model( + id="gpt-5.2-pro", + provider=Provider.OPENAI, + display_name="GPT-5.2 Pro", + streaming=True, + parameter_constraints={ + Parameter.MAX_TOKENS: Range(min=1, max=128000), + TextGenerationParameter.THINKING_BUDGET: Choice( + options=["minimal", "low", "medium", "high", "xhigh"] + ), + TextGenerationParameter.VERBOSITY: Choice( + options=["low", "medium", "high"] + ), + TextGenerationParameter.OUTPUT_SCHEMA: Schema(), + }, + ), + Model( + id="gpt-5.2-chat-latest", + provider=Provider.OPENAI, + display_name="GPT-5.2 Instant", + streaming=True, + parameter_constraints={ + Parameter.TEMPERATURE: Range(min=0.0, max=2.0), + Parameter.MAX_TOKENS: Range(min=1, max=128000), + }, + ), Model( id="gpt-5.1", provider=Provider.OPENAI,