diff --git a/docs/cody/capabilities/supported-models.mdx b/docs/cody/capabilities/supported-models.mdx index 8cb76e65f..db19ed062 100644 --- a/docs/cody/capabilities/supported-models.mdx +++ b/docs/cody/capabilities/supported-models.mdx @@ -15,6 +15,7 @@ Cody supports a variety of cutting-edge large language models for use in chat an | Anthropic | [claude-3 Haiku](https://docs.anthropic.com/claude/docs/models-overview#model-comparison) | ✅ | ✅ | ✅ | | | | | | Anthropic | [claude-3 Sonnet](https://docs.anthropic.com/claude/docs/models-overview#model-comparison) | ✅ | ✅ | ✅ | | | | | | Anthropic | [claude-3.5 Sonnet](https://docs.anthropic.com/claude/docs/models-overview#model-comparison) | ✅ | ✅ | ✅ | | | | | +| Anthropic | [claude-3.5 Sonnet (New)](https://docs.anthropic.com/claude/docs/models-overview#model-comparison) | ✅ | ✅ | ✅ | | | | | | Anthropic | [claude-3 Opus](https://docs.anthropic.com/claude/docs/models-overview#model-comparison) | - | ✅ | ✅ | | | | | | Mistral | [mixtral 8x7b](https://mistral.ai/technology/#models:~:text=of%20use%20cases.-,Mixtral%208x7B,-Currently%20the%20best) | ✅ | ✅ | - | | | | | | Mistral | [mixtral 8x22b](https://mistral.ai/technology/#models:~:text=of%20use%20cases.-,Mixtral%208x7B,-Currently%20the%20best) | ✅ | ✅ | - | | | | | diff --git a/docs/cody/core-concepts/token-limits.mdx b/docs/cody/core-concepts/token-limits.mdx index 4e5d6732c..bba6fa5c7 100644 --- a/docs/cody/core-concepts/token-limits.mdx +++ b/docs/cody/core-concepts/token-limits.mdx @@ -12,20 +12,21 @@ Here's a detailed breakdown of the token limits by model: -| **Model** | **Conversation Context** | **@-mention Context** | **Output** | -| --------------------- | ------------------------ | --------------------- | ---------- | -| gpt-3.5-turbo | 7,000 | shared | 4,000 | -| gpt-4-turbo | 7,000 | shared | 4,000 | -| gpt 4o | 7,000 | shared | 4,000 | -| claude-2.0 | 7,000 | shared | 4,000 | -| claude-2.1 | 7,000 | shared | 4,000 | -| claude-3 Haiku | 7,000 | shared | 4,000 | -| **claude-3 Sonnet** | **15,000** | **30,000** | **4,000** | -| **claude-3.5 Sonnet** | **15,000** | **30,000** | **4,000** | -| mixtral 8x7B | 7,000 | shared | 4,000 | -| mixtral 8x22B | 7,000 | shared | 4,000 | -| Google Gemini 1.5 Flash | 7,000 | shared | 4,000 | -| Google Gemini 1.5 Pro | 7,000 | shared | 4,000 | +| **Model** | **Conversation Context** | **@-mention Context** | **Output** | +| --------------------------- | ------------------------ | --------------------- | ---------- | +| gpt-3.5-turbo | 7,000 | shared | 4,000 | +| gpt-4-turbo | 7,000 | shared | 4,000 | +| gpt 4o | 7,000 | shared | 4,000 | +| claude-2.0 | 7,000 | shared | 4,000 | +| claude-2.1 | 7,000 | shared | 4,000 | +| claude-3 Haiku | 7,000 | shared | 4,000 | +| **claude-3 Sonnet** | **15,000** | **30,000** | **4,000** | +| **claude-3.5 Sonnet** | **15,000** | **30,000** | **4,000** | +| **claude-3.5 Sonnet (New)** | **15,000** | **30,000** | **4,000** | +| mixtral 8x7B | 7,000 | shared | 4,000 | +| mixtral 8x22B | 7,000 | shared | 4,000 | +| Google Gemini 1.5 Flash | 7,000 | shared | 4,000 | +| Google Gemini 1.5 Pro | 7,000 | shared | 4,000 | @@ -43,6 +44,7 @@ Here's a detailed breakdown of the token limits by model: | claude-3 Haiku | 7,000 | shared | 4,000 | | **claude-3 Sonnet** | **15,000** | **30,000** | **4,000** | | **claude-3.5 Sonnet** | **15,000** | **30,000** | **4,000** | +| **claude-3.5 Sonnet (New)** | **15,000** | **30,000** | **4,000** | | **claude-3 Opus** | **15,000** | **30,000** | **4,000** | | **Google Gemini 1.5 Flash** | **15,000** | **30,000** | **4,000** | | **Google Gemini 1.5 Pro** | **15,000** | **30,000** | **4,000** | @@ -50,24 +52,24 @@ Here's a detailed breakdown of the token limits by model: -| **Model** | **Conversation Context** | **@-mention Context** | **Output** | -| ------------------------ | ------------------------ | --------------------- | ---------- | -| gpt-3.5-turbo | 7,000 | shared | 1,000 | -| gpt-4 | 7,000 | shared | 1,000 | -| gpt-4-turbo | 7,000 | shared | 1,000 | -| claude instant | 7,000 | shared | 1,000 | -| claude-2.0 | 7,000 | shared | 1,000 | -| claude-2.1 | 7,000 | shared | 1,000 | -| claude-3 Haiku | 7,000 | shared | 1,000 | -| **claude-3 Sonnet** | **15,000** | **30,000** | **4,000** | -| **claude-3.5 Sonnet^*^** | **15,000** | **30,000** | **4,000** | -| **claude-3 Opus** | **15,000** | **30,000** | **4,000** | -| mixtral 8x7b | 7,000 | shared | 1,000 | +| **Model** | **Conversation Context** | **@-mention Context** | **Output** | +| --------------------------- | ------------------------ | --------------------- | ---------- | +| gpt-3.5-turbo | 7,000 | shared | 1,000 | +| gpt-4 | 7,000 | shared | 1,000 | +| gpt-4-turbo | 7,000 | shared | 1,000 | +| claude instant | 7,000 | shared | 1,000 | +| claude-2.0 | 7,000 | shared | 1,000 | +| claude-2.1 | 7,000 | shared | 1,000 | +| claude-3 Haiku | 7,000 | shared | 1,000 | +| **claude-3 Sonnet** | **15,000** | **30,000** | **4,000** | +| **claude-3.5 Sonnet** | **15,000** | **30,000** | **4,000** | +| **claude-3.5 Sonnet (New)** | **15,000** | **30,000** | **4,000** | +| **claude-3 Opus** | **15,000** | **30,000** | **4,000** | +| mixtral 8x7b | 7,000 | shared | 1,000 | -^*^*Shown are the standard limits. Exact token limits may vary depending on your deployment. Please contact your Sourcegraph representative to learn more.* -For more information on how Cody builds context, see our [documentation here](/cody/core-concepts/context). +For Cody Enterprise, the token limits are the standard limits. Exact token limits may vary depending on your deployment. Please contact your Sourcegraph representative. For more information on how Cody builds context, see our [docs here](/cody/core-concepts/context). ## What is a Context Window?