-
Notifications
You must be signed in to change notification settings - Fork 2.6k
Description
What specific problem does this solve?
GPT-5 and GPT-5 Mini sometimes take long to generate a response. The software Cursor has a solution for this: it features a so-called Fast Mode, which significantly speeds up processing through the OpenAI API. All users are affected by these issues. However, there is a solution: the 'Priority processing' process, which can be activated. Roo Code has not yet implemented this option.
Additional context (optional)
Priority processing:
https://platform.openai.com/docs/guides/priority-processing
Pricing:
https://platform.openai.com/docs/pricing?latest-pricing=priority
Roo Code Task Links (Optional)
No response
Request checklist
- I've searched existing Issues and Discussions for duplicates
- This describes a specific problem with clear impact and context
Interested in implementing this?
- Yes, I'd like to help implement this feature
Implementation requirements
- I understand this needs approval before implementation begins
How should this be solved? (REQUIRED if contributing, optional otherwise)
I believe the best location to enable this option would be in the first menu—under Providers—specifically in the settings available for each model. Currently, users can adjust reasoning and verbosity, and I believe the Priority (fast mode) option should also be included there. Additionally, a clear warning should be displayed indicating that higher costs will apply. Perhaps even with a link to the pricing page.
Pricing:
https://platform.openai.com/docs/pricing?latest-pricing=priority
This option should be available for all providers that offer GPT-5 and GPT-5-mini. In my opinion, these are OpenAI, OpenRouter, and Copilot. If, for technical reasons, it is only possible with the official OpenAI API, I would be completely satisfied with that as well.
How will we know it works? (Acceptance Criteria - REQUIRED if contributing, optional otherwise)
The user should be able to enable the Fast Mode or the original "Priority processing" mode within the provider or model settings.
When the mode is activated, the API response time should noticeably decrease, and the API costs for the same prompt should be higher.
The API processing time should not be as long as when the feature is disabled, and higher costs should occur. This would be a failed test.
Technical considerations (REQUIRED if contributing, optional otherwise)
No response
Trade-offs and risks (REQUIRED if contributing, optional otherwise)
It is important to inform the user that higher costs will occur as a result, otherwise users might become confused or frustrated.
Metadata
Metadata
Assignees
Labels
Type
Projects
Status