diff --git a/backend/utils/llm/conversation_processing.py b/backend/utils/llm/conversation_processing.py index 4da6b5ecbed..83a662caecb 100644 --- a/backend/utils/llm/conversation_processing.py +++ b/backend/utils/llm/conversation_processing.py @@ -257,5 +257,5 @@ def generate_summary_with_prompt(conversation_text: str, prompt: str) -> str: You must output only the summary, no other text. Make sure to be concise and clear. """ - response = llm_mini.invoke(prompt) + response = llm_medium_experiment.invoke(prompt) return response.content \ No newline at end of file