diff --git a/core/llm/llms/CometAPI.ts b/core/llm/llms/CometAPI.ts index 8d5f584e70b..68e91f47af2 100644 --- a/core/llm/llms/CometAPI.ts +++ b/core/llm/llms/CometAPI.ts @@ -1,5 +1,5 @@ -import { LLMOptions } from "../../index.js"; import { allModelProviders } from "@continuedev/llm-info"; +import { LLMOptions } from "../../index.js"; import OpenAI from "./OpenAI.js"; /** @@ -183,6 +183,7 @@ class CometAPI extends OpenAI { */ private static RECOMMENDED_MODELS = [ // GPT series + "gpt-5.1", "gpt-5-chat-latest", "chatgpt-4o-latest", "gpt-5-mini", diff --git a/core/llm/toolSupport.ts b/core/llm/toolSupport.ts index a101779b979..93a0313d81e 100644 --- a/core/llm/toolSupport.ts +++ b/core/llm/toolSupport.ts @@ -392,7 +392,7 @@ export function isRecommendedAgentModel(modelName: string): boolean { [/o[134]/], [/deepseek/, /r1|reasoner/], [/gemini/, /2\.5/, /pro/], - [/gpt-5/], + [/gpt/, /-5|5\.1/], [/claude/, /sonnet/, /3\.7|3-7|-4/], [/claude/, /opus/, /-4/], [/grok-code/], diff --git a/docs/customize/deep-dives/autocomplete.mdx b/docs/customize/deep-dives/autocomplete.mdx index 6648c80e141..d2c2b28bae6 100644 --- a/docs/customize/deep-dives/autocomplete.mdx +++ b/docs/customize/deep-dives/autocomplete.mdx @@ -163,9 +163,9 @@ The `config.json` configuration format offers configuration options through `tab ## Autocomplete FAQs and Troubleshooting in Continue -### I want better completions, should I use GPT-4? +### I want better completions, should I use GPT-5? -Perhaps surprisingly, the answer is no. The models that we suggest for autocomplete are trained with a highly specific prompt format, which allows them to respond to requests for completing code (see examples of these prompts [here](https://github.com/continuedev/continue/blob/main/core/autocomplete/templating/AutocompleteTemplate.ts)). Some of the best commercial models like GPT-4 or Claude are not trained with this prompt format, which means that they won't generate useful completions. Luckily, a huge model is not required for great autocomplete. Most of the state-of-the-art autocomplete models are no more than 10b parameters, and increasing beyond this does not significantly improve performance. +Perhaps surprisingly, the answer is no. The models that we suggest for autocomplete are trained with a highly specific prompt format, which allows them to respond to requests for completing code (see examples of these prompts [here](https://github.com/continuedev/continue/blob/main/core/autocomplete/templating/AutocompleteTemplate.ts)). Some of the best commercial models like GPT-5 or Claude are not trained with this prompt format, which means that they won't generate useful completions. Luckily, a huge model is not required for great autocomplete. Most of the state-of-the-art autocomplete models are no more than 10b parameters, and increasing beyond this does not significantly improve performance. ### Autocomplete Not Working – How to Fix It diff --git a/docs/customize/deep-dives/model-capabilities.mdx b/docs/customize/deep-dives/model-capabilities.mdx index d7b26dc48b2..2e50eb1ab41 100644 --- a/docs/customize/deep-dives/model-capabilities.mdx +++ b/docs/customize/deep-dives/model-capabilities.mdx @@ -171,6 +171,8 @@ This matrix shows which models support tool use and image input capabilities. Co | Model | Tool Use | Image Input | Context Window | | :------------ | -------- | ----------- | -------------- | +| GPT-5.1 | Yes | No | 400k | +| GPT-5 | Yes | No | 400k | | o3 | Yes | No | 128k | | o3-mini | Yes | No | 128k | | GPT-4o | Yes | Yes | 128k | diff --git a/docs/customize/model-roles/chat.mdx b/docs/customize/model-roles/chat.mdx index ee175829fc1..4f92341d2f7 100644 --- a/docs/customize/model-roles/chat.mdx +++ b/docs/customize/model-roles/chat.mdx @@ -76,20 +76,20 @@ If you prefer to use an open-weight model, then the Gemma family of Models from -### GPT-4o from OpenAI +### GPT-5.1 from OpenAI -If you prefer to use a model from [OpenAI](../model-providers/top-level/openai), then we recommend GPT-4o. +If you prefer to use a model from [OpenAI](../model-providers/top-level/openai), then we recommend GPT-5.1. - Add the [OpenAI GPT-4o block](https://hub.continue.dev/openai/gpt-4o) from the hub + Add the [OpenAI GPT-5.1 block](https://hub.continue.dev/openai/gpt-5.1) from the hub ```yaml title="config.yaml" models: - - name: GPT-4o + - name: GPT-5.1 provider: openai - model: '' + model: gpt-5.1 apiKey: ``` diff --git a/gui/src/pages/AddNewModel/configs/models.ts b/gui/src/pages/AddNewModel/configs/models.ts index c9e0c049f56..3d4fe62dc64 100644 --- a/gui/src/pages/AddNewModel/configs/models.ts +++ b/gui/src/pages/AddNewModel/configs/models.ts @@ -1121,6 +1121,18 @@ export const models: { [key: string]: ModelPackage } = { icon: "openai.png", isOpenSource: false, }, + gpt5_1: { + title: "GPT-5.1", + description: "OpenAI's GPT-5.1 model for advanced reasoning and chat", + params: { + model: "gpt-5.1", + contextLength: 400_000, + title: "GPT-5.1", + }, + providerOptions: ["openai"], + icon: "openai.png", + isOpenSource: false, + }, gpt5Codex: { title: "GPT-5 Codex", description: diff --git a/gui/src/pages/AddNewModel/configs/providers.ts b/gui/src/pages/AddNewModel/configs/providers.ts index f988ff80563..436090ec16d 100644 --- a/gui/src/pages/AddNewModel/configs/providers.ts +++ b/gui/src/pages/AddNewModel/configs/providers.ts @@ -111,13 +111,14 @@ export const providers: Partial> = { openai: { title: "OpenAI", provider: "openai", - description: "Use gpt-5, gpt-4, or any other OpenAI model", + description: "Use gpt-5.1, gpt-5, gpt-4, or any other OpenAI model", longDescription: - "Use gpt-5, gpt-4, or any other OpenAI model. See [here](https://openai.com/product#made-for-developers) to obtain an API key.", + "Use gpt-5.1, gpt-5, gpt-4, or any other OpenAI model. See [here](https://openai.com/product#made-for-developers) to obtain an API key.", icon: "openai.png", tags: [ModelProviderTags.RequiresApiKey], packages: [ models.gpt5, + models.gpt5_1, models.gpt5Codex, models.gpt4o, models.gpt4omini, diff --git a/packages/llm-info/src/providers/openai.ts b/packages/llm-info/src/providers/openai.ts index 730495e9c02..516d2c3e59d 100644 --- a/packages/llm-info/src/providers/openai.ts +++ b/packages/llm-info/src/providers/openai.ts @@ -92,6 +92,14 @@ export const OpenAi: ModelProvider = { regex: /gpt-5-codex/, recommendedFor: ["chat"], }, + { + model: "gpt-5.1", + displayName: "GPT-5.1", + contextLength: 400000, + maxCompletionTokens: 128000, + regex: /^gpt-5\.1$/, + recommendedFor: ["chat"], + }, // gpt-4o { model: "gpt-4o",