diff --git a/openevolve/llm/openai.py b/openevolve/llm/openai.py index 7477e5b349..beb5711819 100644 --- a/openevolve/llm/openai.py +++ b/openevolve/llm/openai.py @@ -157,6 +157,7 @@ async def generate_with_context( "temperature": kwargs.get("temperature", self.temperature), "top_p": kwargs.get("top_p", self.top_p), "max_tokens": kwargs.get("max_tokens", self.max_tokens), + "stream": False, # Explicitly request non-streaming responses to avoid SSE issues } # Handle reasoning_effort for open source reasoning models.