From 2c8417b3bf74a68d11364927cb6aef80a159337d Mon Sep 17 00:00:00 2001 From: Maedah Batool Date: Thu, 13 Feb 2025 17:04:12 -0800 Subject: [PATCH] Remove deprecated models --- docs/cody/capabilities/chat.mdx | 2 +- docs/cody/capabilities/supported-models.mdx | 24 ++-- docs/cody/clients/install-eclipse.mdx | 2 +- docs/cody/clients/install-visual-studio.mdx | 2 +- docs/cody/clients/install-vscode.mdx | 8 +- docs/cody/core-concepts/token-limits.mdx | 111 +++++++----------- .../enterprise/completions-configuration.mdx | 2 +- .../cody/enterprise/model-config-examples.mdx | 24 ---- docs/cody/enterprise/model-configuration.mdx | 48 -------- docs/cody/faq.mdx | 3 - docs/cody/prompts-guide.mdx | 2 +- docs/cody/troubleshooting.mdx | 3 - docs/cody/usage-and-pricing.mdx | 20 ++-- docs/pricing/free.mdx | 4 +- .../constants/supportedModelsConfiguration.ts | 42 ------- src/utils/constants/supportedModelsEnt.ts | 42 ------- 16 files changed, 70 insertions(+), 269 deletions(-) diff --git a/docs/cody/capabilities/chat.mdx b/docs/cody/capabilities/chat.mdx index b0e3b6ab0..c532345c7 100644 --- a/docs/cody/capabilities/chat.mdx +++ b/docs/cody/capabilities/chat.mdx @@ -66,7 +66,7 @@ Cody chat can run offline with Ollama. The offline mode does not require you to ![offline-cody-with-ollama](https://storage.googleapis.com/sourcegraph-assets/Docs/cody-offline-ollama.jpg) -You can still switch to your Sourcegraph account whenever you want to use Claude, OpenAI, Gemini, Mixtral, etc. +You can still switch to your Sourcegraph account whenever you want to use Claude, OpenAI, Gemini, etc. ## LLM selection diff --git a/docs/cody/capabilities/supported-models.mdx b/docs/cody/capabilities/supported-models.mdx index fb5439608..be4999d93 100644 --- a/docs/cody/capabilities/supported-models.mdx +++ b/docs/cody/capabilities/supported-models.mdx @@ -8,25 +8,20 @@ Cody supports a variety of cutting-edge large language models for use in chat an | **Provider** | **Model** | **Free** | **Pro** | **Enterprise** | | | | | | :------------ | :-------------------------------------------------------------------------------------------------------------------------------------------- | :----------- | :----------- | :------------- | --- | --- | --- | --- | -| OpenAI | [gpt-3.5 turbo](https://platform.openai.com/docs/models/gpt-3-5-turbo) | ✅ | ✅ | ✅ | | | | | -| OpenAI | [gpt-4](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo#:~:text=to%20Apr%202023-,gpt%2D4,-Currently%20points%20to) | - | - | ✅ | | | | | | OpenAI | [gpt-4 turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo#:~:text=TRAINING%20DATA-,gpt%2D4%2D0125%2Dpreview,-New%20GPT%2D4) | - | ✅ | ✅ | | | | | -| OpenAI | [gpt-4o](https://platform.openai.com/docs/models/gpt-4o) | - | ✅ | ✅ | | | | | -| Anthropic | [claude-3 Haiku](https://docs.anthropic.com/claude/docs/models-overview#model-comparison) | ✅ | ✅ | ✅ | | | | | +| OpenAI | [gpt-4o](https://platform.openai.com/docs/models#gpt-4o) | - | ✅ | ✅ | | | | | +| OpenAI | [gpt-4o-mini](https://platform.openai.com/docs/models#gpt-4o-mini) | ✅ | ✅ | ✅ | | | | | +| OpenAI | [o3-mini-medium](https://openai.com/index/openai-o3-mini/) (experimental) | ✅ | ✅ | ✅ | | | | | +| OpenAI | [o3-mini-high](https://openai.com/index/openai-o3-mini/) (experimental) | - | - | ✅ | | | | | +| OpenAI | [o1](https://platform.openai.com/docs/models#o1) | - | ✅ | ✅ | | | | | | Anthropic | [claude-3.5 Haiku](https://docs.anthropic.com/claude/docs/models-overview#model-comparison) | ✅ | ✅ | ✅ | | | | | -| Anthropic | [claude-3 Sonnet](https://docs.anthropic.com/claude/docs/models-overview#model-comparison) | ✅ | ✅ | ✅ | | | | | | Anthropic | [claude-3.5 Sonnet](https://docs.anthropic.com/claude/docs/models-overview#model-comparison) | ✅ | ✅ | ✅ | | | | | -| Anthropic | [claude-3.5 Sonnet (New)](https://docs.anthropic.com/claude/docs/models-overview#model-comparison) | ✅ | ✅ | ✅ | | | | | -| Anthropic | [claude-3 Opus](https://docs.anthropic.com/claude/docs/models-overview#model-comparison) | - | ✅ | ✅ | | | | | -| Mistral | [mixtral 8x7b](https://mistral.ai/technology/#models:~:text=of%20use%20cases.-,Mixtral%208x7B,-Currently%20the%20best) | ✅ | ✅ | - | | | | | -| Mistral | [mixtral 8x22b](https://mistral.ai/technology/#models:~:text=of%20use%20cases.-,Mixtral%208x7B,-Currently%20the%20best) | ✅ | ✅ | - | | | | | | Ollama | [variety](https://ollama.com/) | experimental | experimental | - | | | | | -| Google Gemini | [1.5 Pro](https://deepmind.google/technologies/gemini/pro/) | ✅ | ✅ | ✅ (Beta) | | | | | -| Google Gemini | [1.5 Flash](https://deepmind.google/technologies/gemini/flash/) | ✅ | ✅ | ✅ (Beta) | | | | | -| Google Gemini | [2.0 Flash Experimental](https://deepmind.google/technologies/gemini/flash/) | ✅ | ✅ | ✅ | | | | | -| | | | | | | | | | +| Google Gemini | [1.5 Pro](https://deepmind.google/technologies/gemini/pro/) | ✅ | ✅ | ✅ (beta) | | | | | +| Google Gemini | [2.0 Flash](https://deepmind.google/technologies/gemini/flash/) | ✅ | ✅ | ✅ | | | | | +| Google Gemini | [2.0 Flash-Lite Preview](https://deepmind.google/technologies/gemini/flash/) (experimental) | ✅ | ✅ | ✅ | | | | | -To use Claude 3 (Opus and Sonnets) models with Cody Enterprise, make sure you've upgraded your Sourcegraph instance to the latest version. +To use Claude 3 Sonnet models with Cody Enterprise, make sure you've upgraded your Sourcegraph instance to the latest version. ## Autocomplete @@ -37,7 +32,6 @@ Cody uses a set of models for autocomplete which are suited for the low latency | Fireworks.ai | [DeepSeek-Coder-V2](https://huggingface.co/deepseek-ai/DeepSeek-Coder-V2-Lite-Instruct) | ✅ | ✅ | ✅ | | | | | | Fireworks.ai | [StarCoder](https://arxiv.org/abs/2305.06161) | - | - | ✅ | | | | | | Anthropic | [claude Instant](https://docs.anthropic.com/claude/docs/models-overview#model-comparison) | - | - | ✅ | | | | | -| Google Gemini (Beta) | [1.5 Flash](https://deepmind.google/technologies/gemini/flash/) | - | - | ✅ | | | | | | Ollama (Experimental) | [variety](https://ollama.com/) | ✅ | ✅ | - | | | | | | | | | | | | | | | diff --git a/docs/cody/clients/install-eclipse.mdx b/docs/cody/clients/install-eclipse.mdx index 2494cb47a..8e55718b7 100644 --- a/docs/cody/clients/install-eclipse.mdx +++ b/docs/cody/clients/install-eclipse.mdx @@ -50,7 +50,7 @@ The chat input field has a default `@-mention` [context chips](#context-retrieva ## LLM selection -Cody offers a variety of large language models (LLMs) to power your chat experience. Cody Free users can access the latest base models from Anthropic, OpenAI, Google, and Mixtral. At the same time, Cody Pro and Enterprise users can access more extended models. +Cody offers a variety of large language models (LLMs) to power your chat experience. Cody Free users can access the latest base models from Anthropic, OpenAI, Google. At the same time, Cody Pro and Enterprise users can access more extended models. Local models are also available through Ollama to Cody Free and Cody Pro users. To use a model in Cody chat, simply download it and run it in Ollama. diff --git a/docs/cody/clients/install-visual-studio.mdx b/docs/cody/clients/install-visual-studio.mdx index 821edd99f..5db5d2a0c 100644 --- a/docs/cody/clients/install-visual-studio.mdx +++ b/docs/cody/clients/install-visual-studio.mdx @@ -43,7 +43,7 @@ The chat input field has a default `@-mention` [context chips](#context-retrieva ## LLM selection -Cody offers a variety of large language models (LLMs) to power your chat experience. Cody Free users can access the latest base models from Anthropic, OpenAI, Google, and Mixtral. At the same time, Cody Pro and Enterprise users can access more extended models. +Cody offers a variety of large language models (LLMs) to power your chat experience. Cody Free users can access the latest base models from Anthropic, OpenAI, Google. At the same time, Cody Pro and Enterprise users can access more extended models. Local models are also available through Ollama to Cody Free and Cody Pro users. To use a model in Cody chat, download it and run it in Ollama. diff --git a/docs/cody/clients/install-vscode.mdx b/docs/cody/clients/install-vscode.mdx index e959beba6..ffe308ae0 100644 --- a/docs/cody/clients/install-vscode.mdx +++ b/docs/cody/clients/install-vscode.mdx @@ -136,7 +136,7 @@ For Edit: - On any file, select some code and a right-click - Select Cody->Edit Code (optionally, you can do this with Opt+K/Alt+K) -- Select the default model available (this is Claude 3 Opus) +- Select the default model available - See the selection of models and click the model you desire. This model will now be the default model going forward on any new edits ### Selecting Context with @-mentions @@ -273,11 +273,11 @@ Users on Cody **Free** and **Pro** can choose from a list of [supported LLM mode ![LLM-models-for-cody-free](https://storage.googleapis.com/sourcegraph-assets/Docs/llm-dropdown-options-2025.png) -Enterprise users get Claude 3 (Opus and Sonnet) as the default LLM models without extra cost. Moreover, Enterprise users can use Claude 3.5 models through Cody Gateway, Anthropic BYOK, Amazon Bedrock (limited availability), and GCP Vertex. +Enterprise users get Claude 3.5 Sonnet as the default LLM models without extra cost. Moreover, Enterprise users can use Claude 3.5 models through Cody Gateway, Anthropic BYOK, Amazon Bedrock (limited availability), and GCP Vertex. For enterprise users on Amazon Bedrock: 3.5 Sonnet is unavailable in `us-west-2` but available in `us-east-1`. Check the current model availability on AWS and your customer's instance location before switching. Provisioned throughput via AWS is not supported for 3.5 Sonnet. -You also get additional capabilities like BYOLLM (Bring Your Own LLM), supporting Single-Tenant and Self Hosted setups for flexible coding environments. Your site administrator determines the LLM, and cannot be changed within the editor. However, Cody Enterprise users when using Cody Gateway have the ability to [configure custom models](/cody/core-concepts/cody-gateway#configuring-custom-models) Anthropic (like Claude 2.0 and Claude Instant), OpenAI (GPT 3.5 and GPT 4) and Google Gemini 1.5 models (Flash and Pro). +You also get additional capabilities like BYOLLM (Bring Your Own LLM), supporting Single-Tenant and Self Hosted setups for flexible coding environments. Your site administrator determines the LLM, and cannot be changed within the editor. However, Cody Enterprise users when using Cody Gateway have the ability to [configure custom models](/cody/core-concepts/cody-gateway#configuring-custom-models) from Anthropic, OpenAI, and Google Gemini. Read more about all the supported LLM models [here](/cody/capabilities/supported-models) @@ -333,7 +333,7 @@ You can use Cody with or without an internet connection. The offline mode does n ![offline-cody-with-ollama](https://storage.googleapis.com/sourcegraph-assets/Docs/cody-offline-ollama.jpg) -You still have the option to switch to your Sourcegraph account whenever you want to use Claude, OpenAI, Gemini, Mixtral, etc. +You still have the option to switch to your Sourcegraph account whenever you want to use Claude, OpenAI, Gemini, etc. ## Experimental models diff --git a/docs/cody/core-concepts/token-limits.mdx b/docs/cody/core-concepts/token-limits.mdx index b83c1ba66..6491aef0a 100644 --- a/docs/cody/core-concepts/token-limits.mdx +++ b/docs/cody/core-concepts/token-limits.mdx @@ -2,7 +2,7 @@

Learn about Cody's token limits and how to manage them.

-For all models, Cody allows up to **4,000 tokens of output**, which is approximately **500-600** lines of code. For Claude 3 Sonnet or Opus models, Cody tracks two separate token limits: +For all models, Cody allows up to **4,000 tokens of output**, which is approximately **500-600** lines of code. For Claude 3 Sonnet models, Cody tracks two separate token limits: - The @-mention context is limited to **30,000 tokens** (~4,000 lines of code) and can be specified using the @-filename syntax. The user explicitly defines this context, which provides specific information to Cody. - Conversation context is limited to **15,000 tokens**, including user questions, system responses, and automatically retrieved context items. Apart from user questions, Cody generates this context automatically. @@ -13,61 +13,54 @@ Here's a detailed breakdown of the token limits by model: -| **Model** | **Conversation Context** | **@-mention Context** | **Output** | -| --------------------------- | ------------------------ | --------------------- | ---------- | -| gpt-3.5-turbo | 7,000 | shared | 4,000 | -| gpt-4-turbo | 7,000 | shared | 4,000 | -| gpt 4o | 7,000 | shared | 4,000 | -| claude-2.0 | 7,000 | shared | 4,000 | -| claude-2.1 | 7,000 | shared | 4,000 | -| claude-3 Haiku | 7,000 | shared | 4,000 | -| claude-3.5 Haiku | 7,000 | shared | 4,000 | -| **claude-3 Sonnet** | **15,000** | **30,000** | **4,000** | -| **claude-3.5 Sonnet** | **15,000** | **30,000** | **4,000** | -| **claude-3.5 Sonnet (New)** | **15,000** | **30,000** | **4,000** | -| mixtral 8x7B | 7,000 | shared | 4,000 | -| mixtral 8x22B | 7,000 | shared | 4,000 | -| Google Gemini 1.5 Flash | 7,000 | shared | 4,000 | -| Google Gemini 1.5 Pro | 7,000 | shared | 4,000 | +| **Model** | **Conversation Context** | **@-mention Context** | **Output** | +| ------------------------------------ | ------------------------ | --------------------- | ---------- | +| gpt-4o-mini | 7,000 | shared | 4,000 | +| gpt-o3-mini-medium | 7,000 | shared | 4,000 | +| claude-3.5 Haiku | 7,000 | shared | 4,000 | +| **claude-3.5 Sonnet (New)** | **15,000** | **30,000** | **4,000** | +| Google Gemini 1.5 Pro | 7,000 | shared | 4,000 | +| Google Gemini 2.0 Flash | 7,000 | shared | 4,000 | +| Google Gemini 2.0 Flash-Lite Preview | 7,000 | shared | 4,000 | -| **Model** | **Conversation Context** | **@-mention Context** | **Output** | -| --------------------------- | ------------------------ | --------------------- | ---------- | -| gpt-3.5-turbo | 7,000 | shared | 4,000 | -| gpt-4 | 7,000 | shared | 4,000 | -| gpt-4-turbo | 7,000 | shared | 4,000 | -| claude instant | 7,000 | shared | 4,000 | -| claude-2.0 | 7,000 | shared | 4,000 | -| claude-2.1 | 7,000 | shared | 4,000 | -| claude-3 Haiku | 7,000 | shared | 4,000 | -| claude-3.5 Haiku | 7,000 | shared | 4,000 | -| **claude-3 Sonnet** | **15,000** | **30,000** | **4,000** | -| **claude-3.5 Sonnet** | **15,000** | **30,000** | **4,000** | -| **claude-3.5 Sonnet (New)** | **15,000** | **30,000** | **4,000** | -| **claude-3 Opus** | **15,000** | **30,000** | **4,000** | -| **Google Gemini 1.5 Flash** | **15,000** | **30,000** | **4,000** | -| **Google Gemini 1.5 Pro** | **15,000** | **30,000** | **4,000** | -| mixtral 8x7b | 7,000 | shared | 4,000 | + +The Pro tier supports the token limits for the LLM models on Free tier, plus: + +| **Model** | **Conversation Context** | **@-mention Context** | **Output** | +| ------------------------------------ | ------------------------ | --------------------- | ---------- | +| gpt-4o-mini | 7,000 | shared | 4,000 | +| gpt-o3-mini-medium | 7,000 | shared | 4,000 | +| gpt-4-turbo | 7,000 | shared | 4,000 | +| gpt-4o | 7,000 | shared | 4,000 | +| o1 | 7,000 | shared | 4,000 | +| claude-3.5 Haiku | 7,000 | shared | 4,000 | +| **claude-3.5 Sonnet (New)** | **15,000** | **30,000** | **4,000** | +| **Google Gemini 1.5 Pro** | **15,000** | **30,000** | **4,000** | +| Google Gemini 2.0 Flash | 7,000 | shared | 4,000 | +| Google Gemini 2.0 Flash-Lite Preview | 7,000 | shared | 4,000 | + -| **Model** | **Conversation Context** | **@-mention Context** | **Output** | -| --------------------------- | ------------------------ | --------------------- | ---------- | -| gpt-3.5-turbo | 7,000 | shared | 1,000 | -| gpt-4 | 7,000 | shared | 1,000 | -| gpt-4-turbo | 7,000 | shared | 1,000 | -| claude instant | 7,000 | shared | 1,000 | -| claude-2.0 | 7,000 | shared | 1,000 | -| claude-2.1 | 7,000 | shared | 1,000 | -| claude-3 Haiku | 7,000 | shared | 1,000 | -| claude-3.5 Haiku | 7,000 | shared | 1,000 | -| **claude-3 Sonnet** | **15,000** | **30,000** | **4,000** | -| **claude-3.5 Sonnet** | **15,000** | **30,000** | **4,000** | -| **claude-3.5 Sonnet (New)** | **15,000** | **30,000** | **4,000** | -| **claude-3 Opus** | **15,000** | **30,000** | **4,000** | -| mixtral 8x7b | 7,000 | shared | 1,000 | + +The Enterprise tier supports the token limits for the LLM models on Free and Pro tier, plus: + +| **Model** | **Conversation Context** | **@-mention Context** | **Output** | +| ------------------------------------ | ------------------------ | --------------------- | ---------- | +| gpt-4o-mini | 7,000 | shared | 1,000 | +| gpt-o3-mini-medium | 7,000 | shared | 1,000 | +| gpt-4-turbo | 7,000 | shared | 1,000 | +| gpt-4o | 7,000 | shared | 1,000 | +| o1 | 7,000 | shared | 1,000 | +| o3-mini-high | 7,000 | shared | 1,000 | +| claude-3.5 Haiku | 7,000 | shared | 1,000 | +| **claude-3.5 Sonnet (New)** | **15,000** | **30,000** | **4,000** | +| Google Gemini 2.0 Flash | 7,000 | shared | 1,000 | +| Google Gemini 2.0 Flash-Lite Preview | 7,000 | shared | 1,000 | + @@ -89,26 +82,6 @@ When a model generates text or code, it does so token by token, predicting the m The output limit helps to keep the generated text focused, concise, and manageable by preventing the model from going off-topic or generating excessively long responses. It also ensures that the output can be efficiently processed and displayed by downstream applications or user interfaces while managing computational resources. -## Current foundation model limits - -Here is a table with the context window sizes and output limits for each of our [supported models](/cody/capabilities/supported-models). - -| **Model** | **Context Window** | **Output Limit** | -| ---------------- | ------------------ | ---------------- | -| gpt-3.5-turbo | 16,385 tokens | 4,096 tokens | -| gpt-4 | 8,192 tokens | 4,096 tokens | -| gpt-4-turbo | 128,000 tokens | 4,096 tokens | -| claude instant | 100,000 tokens | 4,096 tokens | -| claude-2.0 | 100,000 tokens | 4,096 tokens | -| claude-2.1 | 200,000 tokens | 4,096 tokens | -| claude-3 Haiku | 200,000 tokens | 4,096 tokens | -| claude-3.5 Haiku | 200,000 tokens | 4,096 tokens | -| claude-3 Sonnet | 200,000 tokens | 4,096 tokens | -| claude-3 Opus | 200,000 tokens | 4,096 tokens | -| mixtral 8x7b | 32,000 tokens | 4,096 tokens | - -These foundation model limits are the LLM models' inherent limits. For instance, Claude 3 models have a 200K context window compared to 8,192 for GPT-4. - ## Tradeoffs: Size, Accuracy, Latency and Cost So why does Cody not use each model's full available context window? We need to consider a few tradeoffs, namely, context size, retrieval accuracy, latency, and costs. diff --git a/docs/cody/enterprise/completions-configuration.mdx b/docs/cody/enterprise/completions-configuration.mdx index 1d0f7ea56..cdab972ab 100644 --- a/docs/cody/enterprise/completions-configuration.mdx +++ b/docs/cody/enterprise/completions-configuration.mdx @@ -8,7 +8,7 @@ ## Using Sourcegraph Cody Gateway -This is the recommended way to configure Cody Enterprise. It supports all the latest models from Anthropic, OpenAI, Mistral, and more without requiring a separate account or incurring separate charges. You can learn more about these in our [supported models](/cody/capabilities/supported-models) docs. +This is the recommended way to configure Cody Enterprise. It supports all the latest models from Anthropic, OpenAI, and more without requiring a separate account or incurring separate charges. You can learn more about these in our [supported models](/cody/capabilities/supported-models) docs. ## Using your organization's account with a model provider diff --git a/docs/cody/enterprise/model-config-examples.mdx b/docs/cody/enterprise/model-config-examples.mdx index 3276aeda5..d6bf85a94 100644 --- a/docs/cody/enterprise/model-config-examples.mdx +++ b/docs/cody/enterprise/model-config-examples.mdx @@ -144,30 +144,6 @@ Below are configuration examples for setting up various LLM providers using BYOK "maxOutputTokens": 4000 } }, - { - "modelRef": "anthropic::2023-06-01::claude-3-haiku", - "displayName": "Claude 3 Haiku", - "modelName": "claude-3-haiku-20240307", - "capabilities": ["chat"], - "category": "speed", - "status": "stable", - "contextWindow": { - "maxInputTokens": 7000, - "maxOutputTokens": 4000 - } - }, - { - "modelRef": "anthropic::2023-06-01::claude-3-haiku", - "displayName": "Claude 3 Haiku", - "modelName": "claude-3-haiku-20240307", - "capabilities": ["edit", "chat"], - "category": "speed", - "status": "stable", - "contextWindow": { - "maxInputTokens": 7000, - "maxOutputTokens": 4000 - } - } ], "defaultModels": { "chat": "anthropic::2024-10-22::claude-3.5-sonnet", diff --git a/docs/cody/enterprise/model-configuration.mdx b/docs/cody/enterprise/model-configuration.mdx index 59f3c45b7..3a564f7d4 100644 --- a/docs/cody/enterprise/model-configuration.mdx +++ b/docs/cody/enterprise/model-configuration.mdx @@ -352,18 +352,6 @@ The response includes: "maxOutputTokens": 4000 } }, - { - "modelRef": "anthropic::2023-06-01::claude-3-opus", - "displayName": "Claude 3 Opus", - "modelName": "claude-3-opus-20240229", - "capabilities": ["chat"], - "category": "other", - "status": "stable", - "contextWindow": { - "maxInputTokens": 45000, - "maxOutputTokens": 4000 - } - }, { "modelRef": "anthropic::2023-06-01::claude-3-haiku", "displayName": "Claude 3 Haiku", @@ -412,30 +400,6 @@ The response includes: "maxOutputTokens": 4000 } }, - { - "modelRef": "google::v1::gemini-1.5-flash", - "displayName": "Gemini 1.5 Flash", - "modelName": "gemini-1.5-flash", - "capabilities": ["chat"], - "category": "speed", - "status": "stable", - "contextWindow": { - "maxInputTokens": 45000, - "maxOutputTokens": 4000 - } - }, - { - "modelRef": "mistral::v1::mixtral-8x7b-instruct", - "displayName": "Mixtral 8x7B", - "modelName": "accounts/fireworks/models/mixtral-8x7b-instruct", - "capabilities": ["chat"], - "category": "speed", - "status": "stable", - "contextWindow": { - "maxInputTokens": 7000, - "maxOutputTokens": 4000 - } - }, { "modelRef": "openai::2024-02-01::gpt-4o", "displayName": "GPT-4o", @@ -460,18 +424,6 @@ The response includes: "maxOutputTokens": 4000 } }, - { - "modelRef": "openai::2024-02-01::cody-chat-preview-002", - "displayName": "OpenAI o1-mini", - "modelName": "cody-chat-preview-002", - "capabilities": ["chat"], - "category": "accuracy", - "status": "waitlist", - "contextWindow": { - "maxInputTokens": 45000, - "maxOutputTokens": 4000 - } - } ], "defaultModels": { "chat": "anthropic::2024-10-22::claude-3-5-sonnet-latest", diff --git a/docs/cody/faq.mdx b/docs/cody/faq.mdx index ebf2215e2..689da159b 100644 --- a/docs/cody/faq.mdx +++ b/docs/cody/faq.mdx @@ -173,8 +173,6 @@ Once you are signed in make sure to re-enable protection. #### Model Selection -- **o1-preview**: Best for complex reasoning and planning -- **o1-mini**: More reliable for straightforward tasks - **Sonnet 3.5**: Better for tasks requiring longer outputs #### Prompting Strategy @@ -187,7 +185,6 @@ Once you are signed in make sure to re-enable protection. #### Response Time - Start with smaller contexts -- Use o1-mini for faster responses - Consider breaking complex tasks into stages #### Quality diff --git a/docs/cody/prompts-guide.mdx b/docs/cody/prompts-guide.mdx index ab42c5434..bd6af1e8a 100644 --- a/docs/cody/prompts-guide.mdx +++ b/docs/cody/prompts-guide.mdx @@ -97,7 +97,7 @@ You can learn more about context [here](/cody/core-concepts/context). ## Selecting the right LLM -Cody offers a variety of LLMs for both chat and in-line edits by all the leading LLM providers. Each LLM has its strengths and weaknesses, so it is important to select the right one for your use case. For example, Claude 3.5 Sonnet and GPT-4o are powerful for code generation and provide accurate results. However, Gemini 1.5 Flash is a decent choice for cost-effective searches. So, you can always optimize your choice of LLM based on your use case. +Cody offers a variety of LLMs for both chat and in-line edits by all the leading LLM providers. Each LLM has its strengths and weaknesses, so it is important to select the right one for your use case. For example, Claude 3.5 Sonnet and GPT-4o are powerful for code generation and provide accurate results. However, Gemini 2.0 Flash is a decent choice for cost-effective searches. So, you can always optimize your choice of LLM based on your use case. Learn more about all the supported LLMs [here](/cody/capabilities/supported-models). diff --git a/docs/cody/troubleshooting.mdx b/docs/cody/troubleshooting.mdx index 3614434b3..69a3ebcb8 100644 --- a/docs/cody/troubleshooting.mdx +++ b/docs/cody/troubleshooting.mdx @@ -273,7 +273,6 @@ Additionally, Eclipse's built-in Error Log can be used to view any uncaught exce Symptoms: - "Request Failed: Request to... failed with 500 Internal Server Error: context deadline exceeded" -- Occurs with both o1-mini and o1-preview - Happens even with relatively small inputs (~220 lines) Solutions: @@ -299,8 +298,6 @@ Symptoms: Solutions: -- For `o1-preview`: Copy the last line and ask to "continue from last line" -- Switch to `o1-mini` for more reliable complete outputs - Break down complex requests into smaller steps - Consider using Sonnet 3.5 for tasks requiring longer outputs diff --git a/docs/cody/usage-and-pricing.mdx b/docs/cody/usage-and-pricing.mdx index 870411119..dbf4f18ed 100644 --- a/docs/cody/usage-and-pricing.mdx +++ b/docs/cody/usage-and-pricing.mdx @@ -37,20 +37,16 @@ Once your Pro subscription is confirmed, click **My subscription** to manage and Cody Pro, designed for individuals or small teams at **$9 per user per month**, offers an enhanced coding experience beyond the free plan. It provides unlimited autocompletion suggestions plus unlimited chat and prompts. It also uses local repository context to enhance Cody's understanding and accuracy. -Cody Pro uses DeepSeek-Coder-V2 by default for autocomplete. Pro accounts also default to the Claude 3.5 Sonnet (New) model for chat and prompts, but users can switch to other LLM model choices with unlimited usage, including: - -- Claude Instant 1.2 -- Claude 2 -- Claude 3 -- ChatGPT 3.5 Turbo -- GPT-4o -- ChatGPT 4 Turbo Preview -- Mixtral -- Google Gemini 1.5 Pro -- Google Gemini 1.5 Flash +Cody Pro uses DeepSeek-Coder-V2 by default for autocomplete. Pro accounts also default to the Claude 3.5 Sonnet (New) model for chat and prompts, but users can switch to other LLM model choices with unlimited usage. You can refer to the [supported LLM models](/cody/capabilities/supported-models) docs for more information. Support for Cody Pro is available through our Support team via support@sourcegraph.com, ensuring prompt assistance and guidance. +### Enterprise Starter + +Cody Pro users can also switch to the Enterprise Starter plan for **$19 per user per month**. This plan includes all the features of Cody Pro plus a multi-tenant Sourcegraph instance with core features like a fully managed version of Sourcegraph (AI + code search + intent detection with integrated search results, with privately indexed code) through a self-serve flow. + +Read more about the [Enterprise Starter plan](/pricing/enterprise-starter). + There will be high daily limits to catch bad actors and prevent abuse, but under most normal usage, Pro users won't experience these limits. ### Downgrading from Pro to Free @@ -100,7 +96,7 @@ You can access your invoices via the [Cody Dashboard](https://sourcegraph.com/co ## Enterprise -Cody Enterprise is designed for enterprises prioritizing security and administrative controls. We offer either seat-based or token based pricing models, depending on what makes most sense for your organization. You get Claude 3 (Opus and Sonnet 3.5) as the default LLM models without extra cost. You also get additional capabilities like BYOLLM (Bring Your Own LLM), supporting Single-Tenant and Self Hosted setups for flexible coding environments. +Cody Enterprise is designed for enterprises prioritizing security and administrative controls. We offer either seat-based or token based pricing models, depending on what makes most sense for your organization. You get Claude Haiku 3.5 and Claude Sonnet 3.5 as the default LLM models without extra cost. You also get additional capabilities like BYOLLM (Bring Your Own LLM), supporting Single-Tenant and Self Hosted setups for flexible coding environments. Security features include SAML/SSO for enhanced authentication and guardrails to enforce coding standards. Cody Enterprise supports advanced Code Graph context and multi-code host context for a deeper understanding of codebases, especially in complex projects. With 24/5 enhanced support, Cody Enterprise ensures timely assistance. diff --git a/docs/pricing/free.mdx b/docs/pricing/free.mdx index 76c5eeee9..45c5b9364 100644 --- a/docs/pricing/free.mdx +++ b/docs/pricing/free.mdx @@ -12,7 +12,7 @@ The Free plan includes the following features: | ----------------------------------------------------------------------------- | --------------------------------------------------- | ------------------ | ------------------------------------------ | ---------------------- | | Reasonable use autocomplete limits | VS Code, JetBrains IDEs, Visual Studio, and Eclipse | Multi-tenant Cloud | SSO/SAML with basic GitHub, GitLab, Google | Community support only | | Reasonable use chat messages and prompts per month | All popular coding languages | - | - | - | -| Multiple LLM selection (Claude 3.5 Sonnet, Gemini 1.5 Pro and Flash, Mixtral) | Natural language search | - | - | - | +| Multiple LLM selection (Claude 3.5 Sonnet, Gemini 1.5 Pro and Flash) | Natural language search | - | - | - | | Support for local Ollama models | All major codehosts (GitHub, GitLab, Bitbucket) | - | - | - | ## Pricing and billing cycle @@ -27,7 +27,7 @@ The Enterprise Starter plan provides extended usage limits and advanced features | ------------------------ | -------------------------------------------------------------------------------------------------------------------------------------------------------- | --------------------------------------------------------------------------------------------------------------------------------------------------------- | | **Description** | - AI editor assistant for hobbyists or light usage | - AI and search for growing organizations hosted on our cloud | | **Price** | - $0/month
- 1 user | - $19/user/month
- Up to 50 devs | -| **AI features** | - Autocompletions
- 200 chat messages and prompts per month
- Multiple LLM choices for chat
- Connect to local Ollama models | - Code autocomplete and chat
- More powerful LLMs for chat (GPT-4o, Claude 3 Opus)
- Intent detection and integrated search results | +| **AI features** | - Autocompletions
- 200 chat messages and prompts per month
- Multiple LLM choices for chat
- Connect to local Ollama models | - Code autocomplete and chat
- More powerful LLMs for chat (GPT-4o)
- Intent detection and integrated search results | | **Code Search features** | N/A | - Code Search
- Symbol Search | | **Deployment types** | - Multi-tenant Coud | - Multi-tenant Cloud
- Private Workspace
- Privately indexed code (100 repos) | | **Compatibility** | - VS Code, JetBrains IDEs, Visual Studio, and Eclipse
- All popular coding languages
Natural language search
- All major code hosts | - VS Code, JetBrains IDEs, Visual Studio, and Eclipse
- All popular coding languages
Natural language search
- Code hosted on GitHub | diff --git a/src/utils/constants/supportedModelsConfiguration.ts b/src/utils/constants/supportedModelsConfiguration.ts index 45cfebbd1..9ebe665ee 100644 --- a/src/utils/constants/supportedModelsConfiguration.ts +++ b/src/utils/constants/supportedModelsConfiguration.ts @@ -7,14 +7,10 @@ export const chatTableDataConf: any = { { provider: 'Anthropic', model: 'Claude-Instant-1.2', status: 'anthropic/claude-instant-1.2' }, { provider: 'Anthropic', model: 'Claude-2.0', status: 'anthropic/claude-2.0' }, { provider: 'Anthropic', model: 'claude-2.1 (See Note #2)', status: 'anthropic/claude-2.1' }, - { provider: 'Anthropic', model: 'claude-3 Haiku', status: 'anthropic/claude-3-haiku-20240307' }, { provider: 'Anthropic', model: 'claude-3 Sonnet', status: 'anthropic/claude-3-sonnet-20240229' }, - { provider: 'Anthropic', model: 'claude-3 Opus', status: 'anthropic/claude-3-opus-20240229' }, { provider: 'Anthropic', model: 'claude-3.5 Sonnet', status: 'anthropic/claude-3-5-sonnet-20240620' }, { provider: 'Anthropic', model: 'claude-3.5 Sonnet (Latest)', status: 'anthropic/claude-3-5-sonnet-20241022' }, { provider: 'Google', model: 'Gemini 1.5 Pro', status: 'google/gemini-1.5-pro-latest' }, - { provider: 'Google', model: 'Gemini 1.5 Flash', status: 'google/gemini-1.5-flash-latest' }, - { provider: 'Fireworks', model: 'mixtral 8x7b', status: '❌' }, { provider: 'Ollama*', model: 'variety', status: '❌' } ], 'OpenAI': [ @@ -25,14 +21,10 @@ export const chatTableDataConf: any = { { provider: 'Anthropic', model: 'Claude-Instant-1.2', status: '❌' }, { provider: 'Anthropic', model: 'Claude-2.0', status: '❌' }, { provider: 'Anthropic', model: 'claude-2.1 (See Note #2)', status: '❌' }, - { provider: 'Anthropic', model: 'claude-3 Haiku', status: '❌' }, { provider: 'Anthropic', model: 'claude-3 Sonnet', status: '❌' }, - { provider: 'Anthropic', model: 'claude-3 Opus', status: '❌' }, { provider: 'Anthropic', model: 'claude-3.5 Sonnet', status: '❌' }, { provider: 'Anthropic', model: 'claude-3.5 Sonnet (Latest)', status: '❌' }, { provider: 'Google', model: 'Gemini 1.5 Pro', status: '❌' }, - { provider: 'Google', model: 'Gemini 1.5 Flash', status: '❌' }, - { provider: 'Fireworks', model: 'mixtral 8x7b', status: '❌' }, { provider: 'Ollama*', model: 'variety', status: '❌' } ], 'Anthropic': [ @@ -43,14 +35,10 @@ export const chatTableDataConf: any = { { provider: 'Anthropic', model: 'Claude-Instant-1.2', status: 'anthropic/claude-instant-1.2' }, { provider: 'Anthropic', model: 'Claude-2.0', status: 'anthropic/claude-2.0' }, { provider: 'Anthropic', model: 'claude-2.1 (See Note #2)', status: 'anthropic/claude-2.1' }, - { provider: 'Anthropic', model: 'claude-3 Haiku', status: 'anthropic/claude-3-haiku-20240307' }, { provider: 'Anthropic', model: 'claude-3 Sonnet', status: 'anthropic/claude-3-sonnet-20240229' }, - { provider: 'Anthropic', model: 'claude-3 Opus', status: 'anthropic/claude-3-opus-20240229' }, { provider: 'Anthropic', model: 'claude-3.5 Sonnet', status: 'anthropic/claude-3-5-sonnet-20240620' }, { provider: 'Anthropic', model: 'claude-3.5 Sonnet (Latest)', status: 'anthropic/claude-3-5-sonnet-20241022' }, { provider: 'Google', model: 'Gemini 1.5 Pro', status: '❌' }, - { provider: 'Google', model: 'Gemini 1.5 Flash', status: '❌' }, - { provider: 'Fireworks', model: 'mixtral 8x7b', status: '❌' }, { provider: 'Ollama*', model: 'variety', status: '❌' } ], 'Google': [ @@ -61,14 +49,10 @@ export const chatTableDataConf: any = { { provider: 'Anthropic', model: 'Claude-Instant-1.2', status: '❌' }, { provider: 'Anthropic', model: 'Claude-2.0', status: '❌' }, { provider: 'Anthropic', model: 'claude-2.1 (See Note #2)', status: '❌' }, - { provider: 'Anthropic', model: 'claude-3 Haiku', status: '❌' }, { provider: 'Anthropic', model: 'claude-3 Sonnet', status: '❌' }, - { provider: 'Anthropic', model: 'claude-3 Opus', status: '❌' }, { provider: 'Anthropic', model: 'claude-3.5 Sonnet', status: '❌' }, { provider: 'Anthropic', model: 'claude-3.5 Sonnet (Latest)', status: '❌' }, { provider: 'Google', model: 'Gemini 1.5 Pro', status: 'google/gemini-1.5-pro-latest' }, - { provider: 'Google', model: 'Gemini 1.5 Flash', status: 'google/gemini-1.5-flash-latest' }, - { provider: 'Fireworks', model: 'mixtral 8x7b', status: '❌' }, { provider: 'Ollama*', model: 'variety', status: '❌' } ], 'Azure OpenAI': [ @@ -79,14 +63,10 @@ export const chatTableDataConf: any = { { provider: 'Anthropic', model: 'Claude-Instant-1.2', status: '❌' }, { provider: 'Anthropic', model: 'Claude-2.0', status: '❌' }, { provider: 'Anthropic', model: 'claude-2.1 (See Note #2)', status: '❌' }, - { provider: 'Anthropic', model: 'claude-3 Haiku', status: '❌' }, { provider: 'Anthropic', model: 'claude-3 Sonnet', status: '❌' }, - { provider: 'Anthropic', model: 'claude-3 Opus', status: '❌' }, { provider: 'Anthropic', model: 'claude-3.5 Sonnet', status: '❌' }, { provider: 'Anthropic', model: 'claude-3.5 Sonnet (Latest)', status: '❌' }, { provider: 'Google', model: 'Gemini 1.5 Pro', status: '❌' }, - { provider: 'Google', model: 'Gemini 1.5 Flash', status: '❌' }, - { provider: 'Fireworks', model: 'mixtral 8x7b', status: '❌' }, { provider: 'Ollama*', model: 'variety', status: '❌' } ], 'Amazon Bedrock': [ @@ -97,14 +77,10 @@ export const chatTableDataConf: any = { { provider: 'Anthropic', model: 'Claude-Instant-1.2', status: 'anthropic.claude-instant-v1' }, { provider: 'Anthropic', model: 'Claude-2.0', status: 'anthropic.claude-v2' }, { provider: 'Anthropic', model: 'claude-2.1 (See Note #2)', status: 'anthropic.claude-v2:1' }, - { provider: 'Anthropic', model: 'claude-3 Haiku', status: 'anthropic.claude-3-haiku-20240307-v1:0' }, { provider: 'Anthropic', model: 'claude-3 Sonnet', status: 'anthropic.claude-3-sonnet-20240229-v1:0' }, - { provider: 'Anthropic', model: 'claude-3 Opus', status: 'anthropic.claude-3-opus-20240229-v1:0' }, { provider: 'Anthropic', model: 'claude-3.5 Sonnet', status: 'anthropic.claude-3-5-sonnet-20240620-v1:0' }, { provider: 'Anthropic', model: 'claude-3.5 Sonnet (Latest)', status: 'anthropic.claude-3-5-sonnet-20241022-v2:0' }, { provider: 'Google', model: 'Gemini 1.5 Pro', status: '❌' }, - { provider: 'Google', model: 'Gemini 1.5 Flash', status: '❌' }, - { provider: 'Fireworks', model: 'mixtral 8x7b', status: '❌' }, { provider: 'Ollama*', model: 'variety', status: '❌' } ], 'GCP Vertex': [ @@ -115,14 +91,10 @@ export const chatTableDataConf: any = { { provider: 'Anthropic', model: 'Claude-Instant-1.2', status: '❌' }, { provider: 'Anthropic', model: 'Claude-2.0', status: '❌' }, { provider: 'Anthropic', model: 'claude-2.1 (See Note #2)', status: '❌' }, - { provider: 'Anthropic', model: 'claude-3 Haiku', status: 'claude-3-haiku@20240307' }, { provider: 'Anthropic', model: 'claude-3 Sonnet', status: 'claude-3-sonnet@20240229' }, - { provider: 'Anthropic', model: 'claude-3 Opus', status: 'claude-3-opus@20240229' }, { provider: 'Anthropic', model: 'claude-3.5 Sonnet', status: 'claude-3-5-sonnet@20240620' }, { provider: 'Anthropic', model: 'claude-3.5 Sonnet (Latest)', status: 'claude-3-5-sonnet@20241022' }, { provider: 'Google', model: 'Gemini 1.5 Pro', status: '❌' }, - { provider: 'Google', model: 'Gemini 1.5 Flash', status: '❌' }, - { provider: 'Fireworks', model: 'mixtral 8x7b', status: '❌' }, { provider: 'Ollama*', model: 'variety', status: '❌' } ] }; @@ -137,8 +109,6 @@ export const autocompleteTableDataConf: any = { { provider: 'Fireworks', model: 'StarCoder', status: 'fireworks/starcoder' }, { provider: 'Fireworks', model: 'DeepSeek Coder V2', status: 'fireworks/deepseek-coder-v2-lite-base' }, { provider: 'Anthropic', model: 'Claude-Instant-1.2', status: 'anthropic/claude-instant-1.2' }, - { provider: 'Anthropic', model: 'Claude-3 Haiku', status: 'anthropic/claude-3-haiku-20240307' }, - { provider: 'Google', model: 'Gemini 1.5 Flash', status: 'google/gemini-1.5-flash' }, { provider: 'Ollama*', model: 'variety', status: '❌' } ], 'OpenAI': [ @@ -150,8 +120,6 @@ export const autocompleteTableDataConf: any = { { provider: 'Fireworks', model: 'StarCoder', status: '❌' }, { provider: 'Fireworks', model: 'DeepSeek Coder V2', status: '❌' }, { provider: 'Anthropic', model: 'Claude-Instant-1.2', status: '❌' }, - { provider: 'Anthropic', model: 'Claude-3 Haiku', status: '❌' }, - { provider: 'Google', model: 'Gemini 1.5 Flash', status: '❌' }, { provider: 'Ollama*', model: 'variety', status: '❌' } ], 'Anthropic': [ @@ -163,8 +131,6 @@ export const autocompleteTableDataConf: any = { { provider: 'Fireworks', model: 'StarCoder', status: '❌' }, { provider: 'Fireworks', model: 'DeepSeek Coder V2', status: '❌' }, { provider: 'Anthropic', model: 'Claude-Instant-1.2', status: 'claude-instant-1.2' }, - { provider: 'Anthropic', model: 'Claude-3 Haiku', status: 'claude-3-haiku-20240307' }, - { provider: 'Google', model: 'Gemini 1.5 Flash', status: '❌' }, { provider: 'Ollama*', model: 'variety', status: '❌' } ], 'Google': [ @@ -176,8 +142,6 @@ export const autocompleteTableDataConf: any = { { provider: 'Fireworks', model: 'StarCoder', status: '❌' }, { provider: 'Fireworks', model: 'DeepSeek Coder V2', status: '❌' }, { provider: 'Anthropic', model: 'Claude-Instant-1.2', status: '❌' }, - { provider: 'Anthropic', model: 'Claude-3 Haiku', status: '❌' }, - { provider: 'Google', model: 'Gemini 1.5 Flash', status: '❌' }, { provider: 'Ollama*', model: 'variety', status: '❌' } ], 'Azure OpenAI': [ @@ -189,8 +153,6 @@ export const autocompleteTableDataConf: any = { { provider: 'Fireworks', model: 'StarCoder', status: '❌' }, { provider: 'Fireworks', model: 'DeepSeek Coder V2', status: '❌' }, { provider: 'Anthropic', model: 'Claude-Instant-1.2', status: '❌' }, - { provider: 'Anthropic', model: 'Claude-3 Haiku', status: '❌' }, - { provider: 'Google', model: 'Gemini 1.5 Flash', status: '❌' }, { provider: 'Ollama*', model: 'variety', status: '❌' } ], 'Amazon Bedrock': [ @@ -202,8 +164,6 @@ export const autocompleteTableDataConf: any = { { provider: 'Fireworks', model: 'StarCoder', status: '❌' }, { provider: 'Fireworks', model: 'DeepSeek Coder V2', status: '❌' }, { provider: 'Anthropic', model: 'Claude-Instant-1.2', status: 'anthropic.claude-instant-v1' }, - { provider: 'Anthropic', model: 'Claude-3 Haiku', status: 'anthropic.claude-3-haiku-20240307-v1:0' }, - { provider: 'Google', model: 'Gemini 1.5 Flash', status: '❌' }, { provider: 'Ollama*', model: 'variety', status: '❌' } ], 'GCP Vertex': [ @@ -215,8 +175,6 @@ export const autocompleteTableDataConf: any = { { provider: 'Fireworks', model: 'StarCoder', status: '❌' }, { provider: 'Fireworks', model: 'DeepSeek Coder V2', status: '❌' }, { provider: 'Anthropic', model: 'Claude-Instant-1.2', status: '❌' }, - { provider: 'Anthropic', model: 'Claude-3 Haiku', status: 'claude-3-haiku@20240307' }, - { provider: 'Google', model: 'Gemini 1.5 Flash', status: '❌' }, { provider: 'Ollama*', model: 'variety', status: '❌' } ] }; diff --git a/src/utils/constants/supportedModelsEnt.ts b/src/utils/constants/supportedModelsEnt.ts index 860a7851b..15c038f0c 100644 --- a/src/utils/constants/supportedModelsEnt.ts +++ b/src/utils/constants/supportedModelsEnt.ts @@ -23,14 +23,10 @@ export const chatTableDataEnt: any = { { provider: 'Anthropic', model: 'Claude-Instant-1.2', status: '✅' }, { provider: 'Anthropic', model: 'claude-2.0', status: '✅' }, { provider: 'Anthropic', model: 'claude-2.1 (See Note #2)', status: '✅' }, - { provider: 'Anthropic', model: 'claude-3 Haiku', status: '✅ *(5.3.9104 and above)' }, { provider: 'Anthropic', model: 'claude-3 Sonnet', status: '✅ *(5.3.9104 and above)' }, - { provider: 'Anthropic', model: 'claude-3 Opus', status: '✅ *(5.3.9104 and above)' }, { provider: 'Anthropic', model: 'claude-3.5 Sonnet', status: '✅ *(5.5.0 and above)' }, { provider: 'Anthropic', model: 'claude-3.5 Sonnet (Latest)', status: '✅ *(5.9 and above)' }, { provider: 'Google', model: 'Gemini 1.5 Pro', status: '✅ *(5.4.5099 and above)' }, - { provider: 'Google', model: 'Gemini 1.5 Flash', status: '✅ *(5.4.5099 and above)' }, - { provider: 'Fireworks', model: 'mixtral 8x7b', status: '❌' }, { provider: 'Ollama*', model: 'variety', status: '❌' } ], 'OpenAI': [ @@ -41,14 +37,10 @@ export const chatTableDataEnt: any = { { provider: 'Anthropic', model: 'Claude-Instant-1.2', status: '❌' }, { provider: 'Anthropic', model: 'claude-2.0', status: '❌' }, { provider: 'Anthropic', model: 'claude-2.1 (See Note #2)', status: '❌' }, - { provider: 'Anthropic', model: 'claude-3 Haiku', status: '❌' }, { provider: 'Anthropic', model: 'claude-3 Sonnet', status: '❌' }, - { provider: 'Anthropic', model: 'claude-3 Opus', status: '❌' }, { provider: 'Anthropic', model: 'claude-3.5 Sonnet', status: '❌' }, { provider: 'Anthropic', model: 'claude-3.5 Sonnet (Latest)', status: '❌' }, { provider: 'Google', model: 'Gemini 1.5 Pro', status: '❌' }, - { provider: 'Google', model: 'Gemini 1.5 Flash', status: '❌' }, - { provider: 'Fireworks', model: 'mixtral 8x7b', status: '❌' }, { provider: 'Ollama*', model: 'variety', status: '❌' } ], 'Anthropic': [ @@ -59,14 +51,10 @@ export const chatTableDataEnt: any = { { provider: 'Anthropic', model: 'Claude-Instant-1.2', status: '✅' }, { provider: 'Anthropic', model: 'claude-2.0', status: '✅' }, { provider: 'Anthropic', model: 'claude-2.1 (See Note #2)', status: '✅' }, - { provider: 'Anthropic', model: 'claude-3 Haiku', status: '✅ *(5.3.9104 and above)' }, { provider: 'Anthropic', model: 'claude-3 Sonnet', status: '✅ *(5.3.9104 and above)' }, - { provider: 'Anthropic', model: 'claude-3 Opus', status: '✅ *(5.3.9104 and above)' }, { provider: 'Anthropic', model: 'claude-3.5 Sonnet', status: '✅ *(5.5.0 and above)' }, { provider: 'Anthropic', model: 'claude-3.5 Sonnet (Latest)', status: '✅ *(5.5.0 and above)' }, { provider: 'Google', model: 'Gemini 1.5 Pro', status: '❌' }, - { provider: 'Google', model: 'Gemini 1.5 Flash', status: '❌' }, - { provider: 'Fireworks', model: 'mixtral 8x7b', status: '❌' }, { provider: 'Ollama*', model: 'variety', status: '❌' } ], 'Google': [ @@ -77,14 +65,10 @@ export const chatTableDataEnt: any = { { provider: 'Anthropic', model: 'Claude-Instant-1.2', status: '❌' }, { provider: 'Anthropic', model: 'claude-2.0', status: '❌' }, { provider: 'Anthropic', model: 'claude-2.1 (See Note #2)', status: '❌' }, - { provider: 'Anthropic', model: 'claude-3 Haiku', status: '❌' }, { provider: 'Anthropic', model: 'claude-3 Sonnet', status: '❌' }, - { provider: 'Anthropic', model: 'claude-3 Opus', status: '❌' }, { provider: 'Anthropic', model: 'claude-3.5 Sonnet', status: '❌' }, { provider: 'Anthropic', model: 'claude-3.5 Sonnet (Latest)', status: '❌' }, { provider: 'Google', model: 'Gemini 1.5 Pro', status: '✅ *(5.4.5099 and above)' }, - { provider: 'Google', model: 'Gemini 1.5 Flash', status: '✅ *(5.4.5099 and above)' }, - { provider: 'Fireworks', model: 'mixtral 8x7b', status: '❌' }, { provider: 'Ollama*', model: 'variety', status: '❌' } ], 'Azure OpenAI': [ @@ -95,14 +79,10 @@ export const chatTableDataEnt: any = { { provider: 'Anthropic', model: 'Claude-Instant-1.2', status: '❌' }, { provider: 'Anthropic', model: 'claude-2.0', status: '❌' }, { provider: 'Anthropic', model: 'claude-2.1 (See Note #2)', status: '❌' }, - { provider: 'Anthropic', model: 'claude-3 Haiku', status: '❌' }, { provider: 'Anthropic', model: 'claude-3 Sonnet', status: '❌' }, - { provider: 'Anthropic', model: 'claude-3 Opus', status: '❌' }, { provider: 'Anthropic', model: 'claude-3.5 Sonnet', status: '❌' }, { provider: 'Anthropic', model: 'claude-3.5 Sonnet (Latest)', status: '❌' }, { provider: 'Google', model: 'Gemini 1.5 Pro', status: '❌' }, - { provider: 'Google', model: 'Gemini 1.5 Flash', status: '❌' }, - { provider: 'Fireworks', model: 'mixtral 8x7b', status: '❌' }, { provider: 'Ollama*', model: 'variety', status: '❌' } ], 'Amazon Bedrock': [ @@ -113,14 +93,10 @@ export const chatTableDataEnt: any = { { provider: 'Anthropic', model: 'Claude-Instant-1.2', status: '✅' }, { provider: 'Anthropic', model: 'claude-2.0', status: '✅' }, { provider: 'Anthropic', model: 'claude-2.1 (See Note #2)', status: '✅' }, - { provider: 'Anthropic', model: 'claude-3 Haiku', status: '✅ *(5.3.9104 and above)' }, { provider: 'Anthropic', model: 'claude-3 Sonnet', status: '✅ *(5.3.9104 and above)' }, - { provider: 'Anthropic', model: 'claude-3 Opus', status: '✅ *(5.3.9104 and above)' }, { provider: 'Anthropic', model: 'claude-3.5 Sonnet', status: '✅ *(5.5.0 and above)' }, { provider: 'Anthropic', model: 'claude-3.5 Sonnet (Latest)', status: '✅ *(5.5.0 and above)' }, { provider: 'Google', model: 'Gemini 1.5 Pro', status: '❌' }, - { provider: 'Google', model: 'Gemini 1.5 Flash', status: '❌' }, - { provider: 'Fireworks', model: 'mixtral 8x7b', status: '❌' }, { provider: 'Ollama*', model: 'variety', status: '❌' } ], 'GCP Vertex': [ @@ -131,14 +107,10 @@ export const chatTableDataEnt: any = { { provider: 'Anthropic', model: 'Claude-Instant-1.2', status: '❌' }, { provider: 'Anthropic', model: 'claude-2.0', status: '❌' }, { provider: 'Anthropic', model: 'claude-2.1 (See Note #2)', status: '❌' }, - { provider: 'Anthropic', model: 'claude-3 Haiku', status: '✅ *(5.3.9104 and above)' }, { provider: 'Anthropic', model: 'claude-3 Sonnet', status: '✅ *(5.3.9104 and above)' }, - { provider: 'Anthropic', model: 'claude-3 Opus', status: '✅ *(5.3.9104 and above)' }, { provider: 'Anthropic', model: 'claude-3.5 Sonnet', status: '✅ *(5.5.0 and above)' }, { provider: 'Anthropic', model: 'claude-3.5 Sonnet (Latest)', status: '✅ *(5.5.0 and above)' }, { provider: 'Google', model: 'Gemini 1.5 Pro', status: '❌' }, - { provider: 'Google', model: 'Gemini 1.5 Flash', status: '❌' }, - { provider: 'Fireworks', model: 'mixtral 8x7b', status: '❌' }, { provider: 'Ollama*', model: 'variety', status: '❌' } ] }; @@ -153,8 +125,6 @@ export const autocompleteTableDataEnt: any = { { provider: 'Fireworks', model: 'StarCoder', status: '✅' }, { provider: 'Fireworks', model: 'DeepSeek Coder V2', status: '✅' }, { provider: 'Anthropic', model: 'Claude-Instant-1.2', status: '✅' }, - { provider: 'Anthropic', model: 'Claude-3 Haiku', status: '✅ *(5.3.9104 and above)' }, - { provider: 'Google', model: 'Gemini 1.5 Flash', status: '✅ *(5.4.5099 and above)' }, { provider: 'Ollama*', model: 'variety', status: '❌' } ], 'OpenAI': [ @@ -166,8 +136,6 @@ export const autocompleteTableDataEnt: any = { { provider: 'Fireworks', model: 'StarCoder', status: '❌' }, { provider: 'Fireworks', model: 'DeepSeek Coder V2', status: '❌' }, { provider: 'Anthropic', model: 'Claude-Instant-1.2', status: '❌' }, - { provider: 'Anthropic', model: 'Claude-3 Haiku', status: '❌' }, - { provider: 'Google', model: 'Gemini 1.5 Flash', status: '❌' }, { provider: 'Ollama*', model: 'variety', status: '❌' } ], 'Anthropic': [ @@ -179,8 +147,6 @@ export const autocompleteTableDataEnt: any = { { provider: 'Fireworks', model: 'StarCoder', status: '❌' }, { provider: 'Fireworks', model: 'DeepSeek Coder V2', status: '❌' }, { provider: 'Anthropic', model: 'Claude-Instant-1.2', status: '✅' }, - { provider: 'Anthropic', model: 'Claude-3 Haiku', status: '✅ *(5.3.9104 and above)' }, - { provider: 'Google', model: 'Gemini 1.5 Flash', status: '❌' }, { provider: 'Ollama*', model: 'variety', status: '❌' } ], 'Google': [ @@ -192,8 +158,6 @@ export const autocompleteTableDataEnt: any = { { provider: 'Fireworks', model: 'StarCoder', status: '❌' }, { provider: 'Fireworks', model: 'DeepSeek Coder V2', status: '❌' }, { provider: 'Anthropic', model: 'Claude-Instant-1.2', status: '❌' }, - { provider: 'Anthropic', model: 'Claude-3 Haiku', status: '❌' }, - { provider: 'Google', model: 'Gemini 1.5 Flash', status: '❌' }, { provider: 'Ollama*', model: 'variety', status: '❌' } ], 'Azure OpenAI': [ @@ -205,8 +169,6 @@ export const autocompleteTableDataEnt: any = { { provider: 'Fireworks', model: 'StarCoder', status: '❌' }, { provider: 'Fireworks', model: 'DeepSeek Coder V2', status: '❌' }, { provider: 'Anthropic', model: 'Claude-Instant-1.2', status: '❌' }, - { provider: 'Anthropic', model: 'Claude-3 Haiku', status: '❌' }, - { provider: 'Google', model: 'Gemini 1.5 Flash', status: '❌' }, { provider: 'Ollama*', model: 'variety', status: '❌' } ], 'Amazon Bedrock': [ @@ -218,8 +180,6 @@ export const autocompleteTableDataEnt: any = { { provider: 'Fireworks', model: 'StarCoder', status: '❌' }, { provider: 'Fireworks', model: 'DeepSeek Coder V2', status: '❌' }, { provider: 'Anthropic', model: 'Claude-Instant-1.2', status: '✅' }, - { provider: 'Anthropic', model: 'Claude-3 Haiku', status: '✅ *(5.3.9104 and above)' }, - { provider: 'Google', model: 'Gemini 1.5 Flash', status: '❌' }, { provider: 'Ollama*', model: 'variety', status: '❌' } ], 'GCP Vertex': [ @@ -231,8 +191,6 @@ export const autocompleteTableDataEnt: any = { { provider: 'Fireworks', model: 'StarCoder', status: '❌' }, { provider: 'Fireworks', model: 'DeepSeek Coder V2', status: '❌' }, { provider: 'Anthropic', model: 'Claude-Instant-1.2', status: '❌' }, - { provider: 'Anthropic', model: 'Claude-3 Haiku', status: '✅ *(5.5.0 and above)' }, - { provider: 'Google', model: 'Gemini 1.5 Flash', status: '❌' }, { provider: 'Ollama*', model: 'variety', status: '❌' } ] };