From 254517cd8d6669b8da28fbd4f20f25144d528ff3 Mon Sep 17 00:00:00 2001 From: Maedah Batool Date: Mon, 4 Nov 2024 18:33:05 -0800 Subject: [PATCH 1/2] Add new Sonnet 3.5 model to Token Limits page --- docs/cody/core-concepts/token-limits.mdx | 60 ++++++++++++------------ 1 file changed, 31 insertions(+), 29 deletions(-) diff --git a/docs/cody/core-concepts/token-limits.mdx b/docs/cody/core-concepts/token-limits.mdx index 4e5d6732c..bba6fa5c7 100644 --- a/docs/cody/core-concepts/token-limits.mdx +++ b/docs/cody/core-concepts/token-limits.mdx @@ -12,20 +12,21 @@ Here's a detailed breakdown of the token limits by model: -| **Model** | **Conversation Context** | **@-mention Context** | **Output** | -| --------------------- | ------------------------ | --------------------- | ---------- | -| gpt-3.5-turbo | 7,000 | shared | 4,000 | -| gpt-4-turbo | 7,000 | shared | 4,000 | -| gpt 4o | 7,000 | shared | 4,000 | -| claude-2.0 | 7,000 | shared | 4,000 | -| claude-2.1 | 7,000 | shared | 4,000 | -| claude-3 Haiku | 7,000 | shared | 4,000 | -| **claude-3 Sonnet** | **15,000** | **30,000** | **4,000** | -| **claude-3.5 Sonnet** | **15,000** | **30,000** | **4,000** | -| mixtral 8x7B | 7,000 | shared | 4,000 | -| mixtral 8x22B | 7,000 | shared | 4,000 | -| Google Gemini 1.5 Flash | 7,000 | shared | 4,000 | -| Google Gemini 1.5 Pro | 7,000 | shared | 4,000 | +| **Model** | **Conversation Context** | **@-mention Context** | **Output** | +| --------------------------- | ------------------------ | --------------------- | ---------- | +| gpt-3.5-turbo | 7,000 | shared | 4,000 | +| gpt-4-turbo | 7,000 | shared | 4,000 | +| gpt 4o | 7,000 | shared | 4,000 | +| claude-2.0 | 7,000 | shared | 4,000 | +| claude-2.1 | 7,000 | shared | 4,000 | +| claude-3 Haiku | 7,000 | shared | 4,000 | +| **claude-3 Sonnet** | **15,000** | **30,000** | **4,000** | +| **claude-3.5 Sonnet** | **15,000** | **30,000** | **4,000** | +| **claude-3.5 Sonnet (New)** | **15,000** | **30,000** | **4,000** | +| mixtral 8x7B | 7,000 | shared | 4,000 | +| mixtral 8x22B | 7,000 | shared | 4,000 | +| Google Gemini 1.5 Flash | 7,000 | shared | 4,000 | +| Google Gemini 1.5 Pro | 7,000 | shared | 4,000 | @@ -43,6 +44,7 @@ Here's a detailed breakdown of the token limits by model: | claude-3 Haiku | 7,000 | shared | 4,000 | | **claude-3 Sonnet** | **15,000** | **30,000** | **4,000** | | **claude-3.5 Sonnet** | **15,000** | **30,000** | **4,000** | +| **claude-3.5 Sonnet (New)** | **15,000** | **30,000** | **4,000** | | **claude-3 Opus** | **15,000** | **30,000** | **4,000** | | **Google Gemini 1.5 Flash** | **15,000** | **30,000** | **4,000** | | **Google Gemini 1.5 Pro** | **15,000** | **30,000** | **4,000** | @@ -50,24 +52,24 @@ Here's a detailed breakdown of the token limits by model: -| **Model** | **Conversation Context** | **@-mention Context** | **Output** | -| ------------------------ | ------------------------ | --------------------- | ---------- | -| gpt-3.5-turbo | 7,000 | shared | 1,000 | -| gpt-4 | 7,000 | shared | 1,000 | -| gpt-4-turbo | 7,000 | shared | 1,000 | -| claude instant | 7,000 | shared | 1,000 | -| claude-2.0 | 7,000 | shared | 1,000 | -| claude-2.1 | 7,000 | shared | 1,000 | -| claude-3 Haiku | 7,000 | shared | 1,000 | -| **claude-3 Sonnet** | **15,000** | **30,000** | **4,000** | -| **claude-3.5 Sonnet^*^** | **15,000** | **30,000** | **4,000** | -| **claude-3 Opus** | **15,000** | **30,000** | **4,000** | -| mixtral 8x7b | 7,000 | shared | 1,000 | +| **Model** | **Conversation Context** | **@-mention Context** | **Output** | +| --------------------------- | ------------------------ | --------------------- | ---------- | +| gpt-3.5-turbo | 7,000 | shared | 1,000 | +| gpt-4 | 7,000 | shared | 1,000 | +| gpt-4-turbo | 7,000 | shared | 1,000 | +| claude instant | 7,000 | shared | 1,000 | +| claude-2.0 | 7,000 | shared | 1,000 | +| claude-2.1 | 7,000 | shared | 1,000 | +| claude-3 Haiku | 7,000 | shared | 1,000 | +| **claude-3 Sonnet** | **15,000** | **30,000** | **4,000** | +| **claude-3.5 Sonnet** | **15,000** | **30,000** | **4,000** | +| **claude-3.5 Sonnet (New)** | **15,000** | **30,000** | **4,000** | +| **claude-3 Opus** | **15,000** | **30,000** | **4,000** | +| mixtral 8x7b | 7,000 | shared | 1,000 | -^*^*Shown are the standard limits. Exact token limits may vary depending on your deployment. Please contact your Sourcegraph representative to learn more.* -For more information on how Cody builds context, see our [documentation here](/cody/core-concepts/context). +For Cody Enterprise, the token limits are the standard limits. Exact token limits may vary depending on your deployment. Please contact your Sourcegraph representative. For more information on how Cody builds context, see our [docs here](/cody/core-concepts/context). ## What is a Context Window? From feeee467ca522d0a95774b94405c5180632cd9e6 Mon Sep 17 00:00:00 2001 From: Maedah Batool Date: Mon, 4 Nov 2024 18:35:49 -0800 Subject: [PATCH 2/2] Add to LLM list --- docs/cody/capabilities/supported-models.mdx | 1 + 1 file changed, 1 insertion(+) diff --git a/docs/cody/capabilities/supported-models.mdx b/docs/cody/capabilities/supported-models.mdx index 8cb76e65f..db19ed062 100644 --- a/docs/cody/capabilities/supported-models.mdx +++ b/docs/cody/capabilities/supported-models.mdx @@ -15,6 +15,7 @@ Cody supports a variety of cutting-edge large language models for use in chat an | Anthropic | [claude-3 Haiku](https://docs.anthropic.com/claude/docs/models-overview#model-comparison) | ✅ | ✅ | ✅ | | | | | | Anthropic | [claude-3 Sonnet](https://docs.anthropic.com/claude/docs/models-overview#model-comparison) | ✅ | ✅ | ✅ | | | | | | Anthropic | [claude-3.5 Sonnet](https://docs.anthropic.com/claude/docs/models-overview#model-comparison) | ✅ | ✅ | ✅ | | | | | +| Anthropic | [claude-3.5 Sonnet (New)](https://docs.anthropic.com/claude/docs/models-overview#model-comparison) | ✅ | ✅ | ✅ | | | | | | Anthropic | [claude-3 Opus](https://docs.anthropic.com/claude/docs/models-overview#model-comparison) | - | ✅ | ✅ | | | | | | Mistral | [mixtral 8x7b](https://mistral.ai/technology/#models:~:text=of%20use%20cases.-,Mixtral%208x7B,-Currently%20the%20best) | ✅ | ✅ | - | | | | | | Mistral | [mixtral 8x22b](https://mistral.ai/technology/#models:~:text=of%20use%20cases.-,Mixtral%208x7B,-Currently%20the%20best) | ✅ | ✅ | - | | | | |