From 68b0b6f5e51b33b50b33ea4079b9d55ca8569001 Mon Sep 17 00:00:00 2001 From: Ammar Bandukwala Date: Mon, 6 Oct 2025 21:19:14 -0500 Subject: [PATCH] Fix OpenAI reasoning item error in Responses API OpenAI Responses API was failing with error: 'Item rs_* of type reasoning was provided without its required following item' This occurred when using reasoning models (gpt-5, o3, o4-mini) with tool calls in multi-turn conversations. Root cause: When using reasoning models with tool calls, OpenAI requires including 'reasoning.encrypted_content' in the include option to properly preserve reasoning context across conversation steps. Solution: Add include: ['reasoning.encrypted_content'] to OpenAI provider options when reasoningEffort is configured. Refs: https://sdk.vercel.ai/providers/ai-sdk-providers/openai#responses-models --- src/utils/ai/providerOptions.ts | 4 ++++ 1 file changed, 4 insertions(+) diff --git a/src/utils/ai/providerOptions.ts b/src/utils/ai/providerOptions.ts index 48a6df859b..d678da6f67 100644 --- a/src/utils/ai/providerOptions.ts +++ b/src/utils/ai/providerOptions.ts @@ -115,6 +115,10 @@ export function buildProviderOptions( ...(reasoningEffort && { reasoningEffort, reasoningSummary: "detailed", // Enable detailed reasoning summaries + // Include reasoning encrypted content to preserve reasoning context across conversation steps + // Required when using reasoning models (gpt-5, o3, o4-mini) with tool calls + // See: https://sdk.vercel.ai/providers/ai-sdk-providers/openai#responses-models + include: ["reasoning.encrypted_content"], }), // Include previousResponseId for persistence (Responses API) ...(previousResponseId && { previousResponseId }),