diff --git a/core/config/default.ts b/core/config/default.ts index 48aef4656b..3fffb94990 100644 --- a/core/config/default.ts +++ b/core/config/default.ts @@ -6,9 +6,9 @@ import { export const defaultConfig: SerializedContinueConfig = { models: [ { - title: "GPT-4-Turbo (Free Trial)", + title: "GPT-4o (Free Trial)", provider: "free-trial", - model: "gpt-4-turbo", + model: "gpt-4o", systemMessage: "You are an expert software developer. You give helpful and concise responses.", }, @@ -19,13 +19,6 @@ export const defaultConfig: SerializedContinueConfig = { systemMessage: "You are an expert software developer. You give helpful and concise responses. Whenever you write a code block you include the language after the opening ticks.", }, - { - title: "GPT-4o (Free Trial)", - provider: "free-trial", - model: "gpt-4o", - systemMessage: - "You are an expert software developer. You give helpful and concise responses.", - }, { title: "Claude 3 Sonnet (Free Trial)", provider: "free-trial", @@ -50,9 +43,9 @@ export const defaultConfig: SerializedContinueConfig = { export const defaultConfigJetBrains: SerializedContinueConfig = { models: [ { - title: "GPT-4-Turbo (Free Trial)", + title: "GPT-4o (Free Trial)", provider: "free-trial", - model: "gpt-4-turbo", + model: "gpt-4o", systemMessage: "You are an expert software developer. You give helpful and concise responses.", }, @@ -63,13 +56,6 @@ export const defaultConfigJetBrains: SerializedContinueConfig = { systemMessage: "You are an expert software developer. You give helpful and concise responses. Whenever you write a code block you include the language after the opening ticks.", }, - { - title: "GPT-4o (Free Trial)", - provider: "free-trial", - model: "gpt-4o", - systemMessage: - "You are an expert software developer. You give helpful and concise responses.", - }, { title: "Claude 3 Sonnet (Free Trial)", provider: "free-trial", @@ -129,7 +115,6 @@ export const defaultContextProvidersVsCode: ContextProviderWithParams[] = [ { name: "code", params: {} }, { name: "docs", params: {} }, { name: "diff", params: {} }, - { name: "open", params: {} }, { name: "terminal", params: {} }, { name: "problems", params: {} }, { name: "folder", params: {} }, @@ -137,5 +122,7 @@ export const defaultContextProvidersVsCode: ContextProviderWithParams[] = [ ]; export const defaultContextProvidersJetBrains: ContextProviderWithParams[] = [ - { name: "open", params: {} }, + { name: "diff", params: {} }, + { name: "folder", params: {} }, + { name: "codebase", params: {} }, ]; diff --git a/core/config/onboarding.ts b/core/config/onboarding.ts index 49ef177867..4dd94e0c2b 100644 --- a/core/config/onboarding.ts +++ b/core/config/onboarding.ts @@ -7,9 +7,9 @@ export function setupOptimizedMode( ...config, models: [ { - title: "GPT-4-Turbo (Free Trial)", + title: "GPT-4o (Free Trial)", provider: "free-trial", - model: "gpt-4-turbo", + model: "gpt-4o", systemMessage: "You are an expert software developer. You give helpful and concise responses.", }, @@ -20,13 +20,6 @@ export function setupOptimizedMode( systemMessage: "You are an expert software developer. You give helpful and concise responses. Whenever you write a code block you include the language after the opening ticks.", }, - { - title: "GPT-4o (Free Trial)", - provider: "free-trial", - model: "gpt-4o", - systemMessage: - "You are an expert software developer. You give helpful and concise responses.", - }, { title: "Claude 3 Sonnet (Free Trial)", provider: "free-trial", diff --git a/core/llm/llms/FreeTrial.ts b/core/llm/llms/FreeTrial.ts index 6b391bf680..2f290625d1 100644 --- a/core/llm/llms/FreeTrial.ts +++ b/core/llm/llms/FreeTrial.ts @@ -128,9 +128,7 @@ class FreeTrial extends BaseLLM { "llama3-70b", "gpt-3.5-turbo", "gpt-4o", - "gpt-4-turbo", "gemini-1.5-pro-latest", - "claude-3-opus-20240229", "claude-3-sonnet-20240229", "claude-3-haiku-20240307", ]; diff --git a/docs/docs/reference/Model Providers/freetrial.md b/docs/docs/reference/Model Providers/freetrial.md index c58617863e..0d56439aa8 100644 --- a/docs/docs/reference/Model Providers/freetrial.md +++ b/docs/docs/reference/Model Providers/freetrial.md @@ -58,9 +58,9 @@ Groq provides lightning fast inference for open-source LLMs like Llama3, up to t { "models": [ { - "title": "GPT-4-Turbo", + "title": "GPT-4o", "provider": "openai", - "model": "gpt-4-turbo", + "model": "gpt-4o", "apiKey": "YOUR_API_KEY" } ] diff --git a/extensions/vscode/config_schema.json b/extensions/vscode/config_schema.json index ba1c66c5c7..fcbcbbd81c 100644 --- a/extensions/vscode/config_schema.json +++ b/extensions/vscode/config_schema.json @@ -438,14 +438,12 @@ "properties": { "model": { "enum": [ - "gpt-4-turbo", "gpt-4o", "llama3-70b", "gpt-3.5-turbo", "phind-codellama-34b", "gemini-pro", "mistral-8x7b", - "claude-3-opus-20240229", "claude-3-sonnet-20240229", "claude-3-haiku-20240307", "starcoder-7b", diff --git a/extensions/vscode/package.json b/extensions/vscode/package.json index c5a9f2d744..f4285d401c 100644 --- a/extensions/vscode/package.json +++ b/extensions/vscode/package.json @@ -1,7 +1,7 @@ { "name": "continue", "icon": "media/icon.png", - "version": "0.8.29", + "version": "0.9.140", "repository": { "type": "git", "url": "https://github.com/continuedev/continue" diff --git a/extensions/vscode/scripts/prepackage.js b/extensions/vscode/scripts/prepackage.js index 323eda7578..32a9e7a5b4 100644 --- a/extensions/vscode/scripts/prepackage.js +++ b/extensions/vscode/scripts/prepackage.js @@ -486,7 +486,7 @@ function validateFilesPresent() { `bin/napi-v3/${os}/${arch}/onnxruntime_binding.node`, `bin/napi-v3/${os}/${arch}/${ os === "darwin" - ? "libonnxruntime.1.17.3.dylib" + ? "libonnxruntime.1.14.0.dylib" : os === "linux" ? "libonnxruntime.so.1.14.0" : "onnxruntime.dll" diff --git a/gui/src/components/Layout.tsx b/gui/src/components/Layout.tsx index efc53da7df..cab8767eb8 100644 --- a/gui/src/components/Layout.tsx +++ b/gui/src/components/Layout.tsx @@ -26,6 +26,7 @@ import { isJetBrains, postToIde } from "../util/ide"; import { getLocalStorage } from "../util/localStorage"; import HeaderButtonWithText from "./HeaderButtonWithText"; import TextDialog from "./dialogs"; +import { ftl } from "./dialogs/FTCDialog"; import IndexingProgressBar from "./loaders/IndexingProgressBar"; import ProgressBar from "./loaders/ProgressBar"; import ModelSelect from "./modelSelection/ModelSelect"; @@ -267,12 +268,10 @@ const Layout = () => { )} */} {indexingState.status !== "indexing" && // Would take up too much space together with indexing progress - defaultModel?.provider === "free-trial" && - (location.pathname === "/settings" || - parseInt(localStorage.getItem("ftc") || "0") >= 10) && ( + defaultModel?.provider === "free-trial" && ( )} diff --git a/gui/src/components/dialogs/FTCDialog.tsx b/gui/src/components/dialogs/FTCDialog.tsx index 8a25e6c7ab..8990c30267 100644 --- a/gui/src/components/dialogs/FTCDialog.tsx +++ b/gui/src/components/dialogs/FTCDialog.tsx @@ -14,6 +14,14 @@ const GridDiv = styled.div` align-items: center; `; +export const ftl = () => { + const ftc = parseInt(localStorage.getItem("ftc")); + if (ftc && ftc > 52) { + return 100; + } + return 50; +}; + function FTCDialog() { const navigate = useNavigate(); const [apiKey, setApiKey] = React.useState(""); @@ -23,9 +31,9 @@ function FTCDialog() {

Free Trial Limit Reached

- You've reached the free trial limit of 100 free inputs. To keep using - Continue, you can either use your own API key, or use a local LLM. To - read more about the options, see our{" "} + You've reached the free trial limit of {ftl()} free inputs. To keep + using Continue, you can either use your own API key, or use a local LLM. + To read more about the options, see our{" "} = 100) { + if (u >= ftl()) { dispatch(setShowDialog(true)); dispatch(setDialogMessage()); posthog?.capture("ftc_reached"); diff --git a/gui/src/util/modelData.ts b/gui/src/util/modelData.ts index a554a715e7..a7797bec7b 100644 --- a/gui/src/util/modelData.ts +++ b/gui/src/util/modelData.ts @@ -546,7 +546,7 @@ const gpt4turbo: ModelPackage = { contextLength: 128_000, title: "GPT-4 Turbo", }, - providerOptions: ["openai", "freetrial"], + providerOptions: ["openai"], icon: "openai.png", }; @@ -1039,7 +1039,6 @@ After it's up and running, you can start using Continue.`, { ...claude3Haiku, title: "Claude 3 Haiku (trial)" }, { ...gemini15Pro, title: "Gemini 1.5 Pro (trial)" }, { ...gpt4o, title: "GPT-4o (trial)" }, - { ...gpt4turbo, title: "GPT-4-Turbo (trial)" }, { ...gpt35turbo, title: "GPT-3.5-Turbo (trial)" }, { ...AUTODETECT,