Skip to content

[BUG] Context Condensing Regression: No dedicated model selector, using active model causes high cost/errors, and switching models breaks continuation #11785

@cmpdev23

Description

@cmpdev23

Problem (one or two sentences)

The option to select a dedicated model for context condensing has been removed from settings. The current workaround (switching models before condensing) breaks the conversation and causes API errors when switching back to the original model, making context condensing effectively unusable.

Context (who is affected and when)

Affects all users who use expensive/slow models (e.g., Claude Opus) for coding and previously relied on a dedicated cheaper model (e.g., Gemini Flash) for context condensing. This impacts anyone working on long tasks where context needs to be condensed regularly, making them choose between high cost or a broken workflow.

Reproduction steps

  1. Open Roo Code settings — notice the option to select a dedicated model for context condensing is no longer present (it was available in a previous version).
    1. Start a long conversation with an expensive model (e.g., Claude Opus 4.6).
    1. When the context gets long, click the "Condense context" button in the chat.
    1. Observe that the condensation runs using the currently active model (Opus 4.6), which is slow and expensive for this simple task.
    1. Workaround attempt: Before condensing, switch to a cheaper/faster model (e.g., Gemini Flash). Condense the context. Then switch back to the original model (Opus 4.6).
    1. Observe that after switching back, the model is no longer able to continue the conversation — it produces only API errors.

Expected result

A dedicated model selector for context condensing should be available in settings (as it was before), allowing users to pick a fast and cheap model (e.g., Gemini Flash) for condensing, independently of the main coding model. After condensing, the original model should seamlessly continue the conversation without errors.

Actual result

  1. The dedicated model selector for context condensing is gone from settings.2. Condensing now always uses the currently active model, making it slow and costly when using a powerful model like Opus.3. If the user manually switches to a cheaper model to condense, then switches back to the original model, the conversation breaks entirely — the model can no longer continue and only returns API errors (likely a system prompt / cross-model configuration conflict).

Variations tried (optional)

  • Tried condensing with the active model (Opus): works but is very slow and expensive.
    • Tried switching to Gemini Flash before condensing, then switching back to Opus: condensing succeeds, but the conversation is broken afterward — only API errors occur when continuing with Opus.
      • The old behavior (dedicated model selector in settings) worked perfectly and this regression makes context condensing practically unusable for users who rely on affordable models for this task.

App Version

latest

API Provider (optional)

None

Model Used (optional)

Claude Opus (main model), Gemini Flash (attempted condensing model)

Roo Code Task Links (optional)

No response

Relevant logs or errors (optional)

Metadata

Metadata

Assignees

No one assigned

    Labels

    bugSomething isn't working

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions