Skip to content

Support "Priority processing" for GPT-5 and GPT-5-mini #7450

@XLeukon

Description

@XLeukon

What specific problem does this solve?

GPT-5 and GPT-5 Mini sometimes take long to generate a response. The software Cursor has a solution for this: it features a so-called Fast Mode, which significantly speeds up processing through the OpenAI API. All users are affected by these issues. However, there is a solution: the 'Priority processing' process, which can be activated. Roo Code has not yet implemented this option.

Additional context (optional)

Priority processing:
https://platform.openai.com/docs/guides/priority-processing
Pricing:
https://platform.openai.com/docs/pricing?latest-pricing=priority

Roo Code Task Links (Optional)

No response

Request checklist

  • I've searched existing Issues and Discussions for duplicates
  • This describes a specific problem with clear impact and context

Interested in implementing this?

  • Yes, I'd like to help implement this feature

Implementation requirements

  • I understand this needs approval before implementation begins

How should this be solved? (REQUIRED if contributing, optional otherwise)

I believe the best location to enable this option would be in the first menu—under Providers—specifically in the settings available for each model. Currently, users can adjust reasoning and verbosity, and I believe the Priority (fast mode) option should also be included there. Additionally, a clear warning should be displayed indicating that higher costs will apply. Perhaps even with a link to the pricing page.
Pricing:
https://platform.openai.com/docs/pricing?latest-pricing=priority

This option should be available for all providers that offer GPT-5 and GPT-5-mini. In my opinion, these are OpenAI, OpenRouter, and Copilot. If, for technical reasons, it is only possible with the official OpenAI API, I would be completely satisfied with that as well.

How will we know it works? (Acceptance Criteria - REQUIRED if contributing, optional otherwise)

The user should be able to enable the Fast Mode or the original "Priority processing" mode within the provider or model settings.

When the mode is activated, the API response time should noticeably decrease, and the API costs for the same prompt should be higher.

The API processing time should not be as long as when the feature is disabled, and higher costs should occur. This would be a failed test.

Technical considerations (REQUIRED if contributing, optional otherwise)

No response

Trade-offs and risks (REQUIRED if contributing, optional otherwise)

It is important to inform the user that higher costs will occur as a result, otherwise users might become confused or frustrated.

Metadata

Metadata

Assignees

No one assigned

    Labels

    Issue - Needs ScopingValid, but needs effort estimate or design input before work can start.enhancementNew feature or requestproposal

    Type

    No type

    Projects

    Status

    Done

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions