What version of Codex CLI is running?
0.124.0
What subscription do you have?
GPT-5.5 is marketed as a Codex model with 400K context and tuned long-running behavior, but it appears not to support /responses/compact/native compaction. As a result, every GPT-5.5 Codex session is effectively non-recoverable once it approaches the context limit. This makes GPT-5.5 unusable for long-running Codex work despite being positioned as the recommended Codex model.
Which model were you using?
gpt-5.5
What platform is your computer?
Microsoft Windows NT 10.0.26200.0 x64
What terminal emulator and version are you using (if applicable)?
No response
What issue are you seeing?
GPT-5.5 is marketed as a Codex model with 400K context and tuned long-running behavior,
but it appears not to support /responses/compact/native compaction.
As a result, every GPT-5.5 Codex session is effectively non-recoverable once it approaches 272k tokens context.
This makes GPT-5.5 unusable for long-running Codex work despite being positioned as the recommended Codex model.
What steps can reproduce the bug?
Environment
- Codex CLI version:
<paste codex --version here>
- OS:
<Windows / macOS / Linux>
- Auth mode: ChatGPT login / Codex subscription
- Plan: Pro
- Model:
gpt-5.5
- Also reproducible when
gpt-5.5 is used as a subagent from a gpt-5.4 parent session.
Config
model = "gpt-5.5"
model_reasoning_effort = "xhigh"
# I tried both the default settings and explicit lower limits.
# The failure still happens once the session gets close to the effective context limit.
model_auto_compact_token_limit = 180000
tool_output_token_limit = 8000
Reproduction steps
Start a fresh Codex session using gpt-5.5 xhigh.
Work on a non-trivial repository task that causes the session context to grow.
For example:
- ask Codex to inspect multiple source and test files;
- run test/build commands with tool output;
- continue several implementation/debugging turns until the session reaches roughly 200k–220k tokens.
- Wait for Codex to trigger remote compaction automatically, or continue the task until compaction is required.
- The remote compaction task fails.
After the first failure, continuing the same session usually keeps failing because the same oversized conversation history is reused. The session effectively becomes unrecoverable.
error belike:
Error running remote compact task:
{"error":{"message":"invalid_request_error: 请求格式非法,请检查请求结构是否符合 API 规范", "type":"invalid_request_error", "code":"invalid_request_error"}}
or
Error running remote compact task:
unexpected status 503 Service Unavailable:
所有供应商暂时不可用
url: /v1/responses/compact
### What is the expected behavior?
_No response_
### Additional information
_No response_
What version of Codex CLI is running?
0.124.0
What subscription do you have?
GPT-5.5 is marketed as a Codex model with 400K context and tuned long-running behavior, but it appears not to support /responses/compact/native compaction. As a result, every GPT-5.5 Codex session is effectively non-recoverable once it approaches the context limit. This makes GPT-5.5 unusable for long-running Codex work despite being positioned as the recommended Codex model.
Which model were you using?
gpt-5.5
What platform is your computer?
Microsoft Windows NT 10.0.26200.0 x64
What terminal emulator and version are you using (if applicable)?
No response
What issue are you seeing?
GPT-5.5 is marketed as a Codex model with 400K context and tuned long-running behavior,
but it appears not to support /responses/compact/native compaction.
As a result, every GPT-5.5 Codex session is effectively non-recoverable once it approaches 272k tokens context.
This makes GPT-5.5 unusable for long-running Codex work despite being positioned as the recommended Codex model.
What steps can reproduce the bug?
Environment
<paste codex --version here><Windows / macOS / Linux>gpt-5.5gpt-5.5is used as a subagent from agpt-5.4parent session.Config