Ollama-compatible API proxy that routes VS Copilot Chat requests to the OpenCode Go subscription API.
When you connect VS Copilot to this proxy, you get access to 14 Go models — including DeepSeek, Qwen, Kimi, MiniMax, GLM, and MiMo — with full agent mode (tool calling) support.
Tested with Visual Studio 2026
-
Install Bun
-
Set your API key
# .env OPENCODE_API_KEY=your-key
-
Run the proxy
bun run start # or on Windows: start.cmd -
Configure VS Code — set
github.copilot.chat.advanced.debug.overrideProxyUrltohttp://localhost:11434andoverrideEngineUrltohttp://localhost:11434/v1
You can register the proxy as an Ollama instance using the project port:
// settings.json
"ollama.endpoint": "http://localhost:11434"| Endpoint | Format | Used By |
|---|---|---|
/api/tags |
Ollama | Copilot model discovery |
/api/chat |
Ollama | Legacy chat |
/api/generate |
Ollama | Legacy completion |
/api/show |
Ollama | Model info |
/v1/chat/completions |
OpenAI | Copilot chat (primary) |
/v1/models |
OpenAI | Model listing |
All models support tool calling (agent mode). Vision-capable models are listed below.
| Model | Vision |
|---|---|
| DeepSeek V4 Flash | |
| Qwen3.5 Plus | ✓ |
| Qwen3.6 Plus | ✓ |
| MiniMax M2.5 | |
| MiniMax M2.7 | |
| Kimi K2.5 | ✓ |
| Kimi K2.6 | ✓ |
| GLM-5 | |
| GLM-5.1 | |
| MiMo V2 Omni | ✓ |
| MiMo V2.5 | ✓ |
| MiMo V2 Pro | ✓ |
| MiMo V2.5 Pro | ✓ |
| Variable | Default |
|---|---|
OPENCODE_API_KEY |
(required) |
SERVER_PORT |
11434 |
SERVER_HOST |
127.0.0.1 |
DEFAULT_MODEL |
deepseek-v4-flash |
- VS Copilot sends model requests in Ollama format (
/api/tags) - Proxy forwards chat to the OpenCode Go API (
/zen/go/v1/chat/completions) - Responses are streamed back in SSE format with proper tool call normalization
- DeepSeek
reasoning_contentis cached and re-injected since VS strips non-standard OpenAI fields