Description
Bad Request: Requested token count exceeds the model's maximum context length of 202752 tokens. You requested a total of 202941 tokens: 170941 tokens from the input messages and 32000 tokens for the completion.
Please reduce the number of tokens in the input messages or the completion to fit within the limit.
That's what sglang is returning
Plugins
No response
OpenCode version
1.14.30
Steps to reproduce
- exceed context length in open code on an sglang (openai compatible). open code does not recover.
Screenshot and/or share link
No response
Operating System
No response
Terminal
No response
Description
Bad Request: Requested token count exceeds the model's maximum context length of 202752 tokens. You requested a total of 202941 tokens: 170941 tokens from the input messages and 32000 tokens for the completion.
Please reduce the number of tokens in the input messages or the completion to fit within the limit.
That's what sglang is returning
Plugins
No response
OpenCode version
1.14.30
Steps to reproduce
Screenshot and/or share link
No response
Operating System
No response
Terminal
No response