export ANTHROPIC_BASE_URL="http://localhost:4000"
export ANTHROPIC_AUTH_TOKEN="sk-1234"
npm install -g @anthropic-ai/claude-code@latest
export LITELLM_API_KEY="sk-1234"
npm install -g @openai/codex@native
model = "gemini-2.5-pro"
model_provider = "litellm"
[model_providers.litellm]
name = "LiteLLM"
base_url = "http://localhost:4000/v1"
env_key = "LITELLM_API_KEY"
wire_api = "chat"
cat $HOME/.codex/config.toml
export GOOGLE_GEMINI_BASE_URL="http://localhost:4000"
export GEMINI_API_KEY="sk-1234"
npm install -g @ai-codespark/gemini-cli@latest
{
"selectedAuthType": "gemini-api-key",
"theme": "ANSI"
}
cat $HOME/.gemini/settings.json
export SILICONFLOW_BASE_URL="http://localhost:4000"
export SILICONFLOW_API_KEY="sk-1234"
export GEMINI_MODEL=gemini-2.5-pro
npm install -g @gen-cli/gen-cli@latest
{
"selectedAuthType": "siliconflow-api-key"
}
cat ~/.gen-cli/settings.json
export OPENAI_BASE_URL="http://localhost:4000"
export OPENAI_API_KEY="sk-1234"
export OPENAI_MODEL="gemini-2.5-pro"
npm install -g @qwen-code/qwen-code@latest
{
"selectedAuthType": "openai"
}
cat ~/.qwen/settings.json
uv venv
uv sync --all-extras
source .venv/bin/activate
{
"default_provider": "openai",
"max_steps": 20,
"enable_lakeview": false,
"model_providers": {
"openai": {
"api_key": "sk-1234",
"base_url": "http://localhost:4000",
"model": "gemini-2.5-pro",
"max_tokens": 120000,
"temperature": 0.5,
"top_p": 1,
"max_retries": 10
}
}
}
cat trae_config.json