diff --git a/docs/cody/clients/install-vscode.mdx b/docs/cody/clients/install-vscode.mdx index 5172bac4f..f9aa58ccf 100644 --- a/docs/cody/clients/install-vscode.mdx +++ b/docs/cody/clients/install-vscode.mdx @@ -366,30 +366,39 @@ Once configured, and VS Code has been restarted, you can select the configured m Example VS Code user settings JSON configuration: ```json -{ - "cody.dev.models": [ - // Google (e.g. Gemini 1.5 Pro) +{ + "cody.dev.models": [ { "provider": "google", - "model": "gemini-1.5-pro-latest", - "tokens": 1000000, - "apiKey": "xyz" + "model": "gemini-2.0-flash-exp", + "inputTokens": 1048576, + "outputTokens": 8192, + "apiKey": "", + "options": { + "temperature": 0.0 + } }, - // Groq (e.g. llama2 70b) { "provider": "groq", "model": "llama2-70b-4096", - "tokens": 4096, - "apiKey": "xyz" + "inputTokens": 4000, + "outputTokens": 4000, + "apiKey": "", + "options": { + "temperature": 0.0 + } }, - // OpenAI & OpenAI-compatible APIs { - "provider": "openai", // keep groq as provider + "provider": "openai", "model": "some-model-id", - "apiKey": "xyz", + "inputTokens": 32000, + "outputTokens": 4000, + "apiKey": "", + "options": { + "temperature": 0.0 + }, "apiEndpoint": "https://host.domain/path" }, - // Ollama (remote) { "provider": "ollama", "model": "some-model-id",