diff --git a/core/autocomplete/templates.ts b/core/autocomplete/templates.ts index 4de43739ab..c0177bb12f 100644 --- a/core/autocomplete/templates.ts +++ b/core/autocomplete/templates.ts @@ -155,11 +155,6 @@ const deepseekFimTemplate: AutocompleteTemplate = { }, }; -const deepseekFimTemplateWrongPipeChar: AutocompleteTemplate = { - template: "<|fim▁begin|>{{{prefix}}}<|fim▁hole|>{{{suffix}}}<|fim▁end|>", - completionOptions: { stop: ["<|fim▁begin|>", "<|fim▁hole|>", "<|fim▁end|>"] }, -}; - const gptAutocompleteTemplate: AutocompleteTemplate = { template: `\`\`\` {{{prefix}}}[BLANK]{{{suffix}}} diff --git a/core/llm/llms/FreeTrial.ts b/core/llm/llms/FreeTrial.ts index b2db6a5a17..e4c1674e63 100644 --- a/core/llm/llms/FreeTrial.ts +++ b/core/llm/llms/FreeTrial.ts @@ -165,7 +165,8 @@ class FreeTrial extends BaseLLM { return [ "codestral-latest", "claude-3-5-sonnet-20240620", - "llama3-70b", + "llama3.1-405b", + "llama3.1-70b", "gpt-4o", "gpt-3.5-turbo", "claude-3-haiku-20240307", diff --git a/core/llm/llms/Groq.ts b/core/llm/llms/Groq.ts index 0aa5b56365..27c51e1e33 100644 --- a/core/llm/llms/Groq.ts +++ b/core/llm/llms/Groq.ts @@ -14,6 +14,9 @@ class Groq extends OpenAI { gemma: "gemma-7b-it", "llama3-8b": "llama3-8b-8192", "llama3-70b": "llama3-70b-8192", + "llama3.1-8b": "llama-3.1-8b-instant", + "llama3.1-70b": "llama-3.1-70b-versatile", + "llama3.1-405b": "llama-3.1-405b-reasoning", }; protected _convertModelName(model: string): string { return Groq.modelConversion[model] ?? model; diff --git a/core/llm/llms/Ollama.ts b/core/llm/llms/Ollama.ts index e24a489f0c..8da67a14e9 100644 --- a/core/llm/llms/Ollama.ts +++ b/core/llm/llms/Ollama.ts @@ -88,6 +88,9 @@ class Ollama extends BaseLLM { "codellama-70b": "codellama:70b", "llama3-8b": "llama3:8b", "llama3-70b": "llama3:70b", + "llama3.1-8b": "llama3.1:8b", + "llama3.1-70b": "llama3.1:70b", + "llama3.1-405b": "llama3.1:405b", "phi-2": "phi:2.7b", "phind-codellama-34b": "phind-codellama:34b-v2", "wizardcoder-7b": "wizardcoder:7b-python", diff --git a/core/llm/llms/Replicate.ts b/core/llm/llms/Replicate.ts index 7cfa6e8901..5dde240148 100644 --- a/core/llm/llms/Replicate.ts +++ b/core/llm/llms/Replicate.ts @@ -18,6 +18,9 @@ class Replicate extends BaseLLM { "llama2-13b": "meta/llama-2-13b-chat" as any, "llama3-8b": "meta/meta-llama-3-8b-instruct" as any, "llama3-70b": "meta/meta-llama-3-70b-instruct" as any, + "llama3.1-8b": "meta/meta-llama-3.1-8b-instruct" as any, + "llama3.1-70b": "meta/meta-llama-3.1-70b-instruct" as any, + "llama3.1-405b": "meta/meta-llama-3.1-405b-instruct" as any, "zephyr-7b": "nateraw/zephyr-7b-beta:b79f33de5c6c4e34087d44eaea4a9d98ce5d3f3a09522f7328eea0685003a931", "mistral-7b": diff --git a/core/llm/llms/Together.ts b/core/llm/llms/Together.ts index 277389c770..eaa1442771 100644 --- a/core/llm/llms/Together.ts +++ b/core/llm/llms/Together.ts @@ -19,6 +19,9 @@ class Together extends OpenAI { "codellama-70b": "codellama/CodeLlama-70b-Instruct-hf", "llama3-8b": "meta-llama/Llama-3-8b-chat-hf", "llama3-70b": "meta-llama/Llama-3-70b-chat-hf", + "llama3.1-8b": "meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo", + "llama3.1-70b": "meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo", + "llama3.1-405b": "meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo", "llama2-7b": "togethercomputer/llama-2-7b-chat", "llama2-13b": "togethercomputer/llama-2-13b-chat", "llama2-70b": "togethercomputer/llama-2-70b-chat", diff --git a/docs/static/schemas/config.json b/docs/static/schemas/config.json index ac7461d8fd..0512f8ac0d 100644 --- a/docs/static/schemas/config.json +++ b/docs/static/schemas/config.json @@ -513,6 +513,9 @@ "anyOf": [ { "enum": [ + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "llama3-8b", "llama3-70b", "codellama-7b", @@ -546,7 +549,8 @@ "enum": [ "gpt-4o", "codestral-latest", - "llama3-70b", + "llama3.1-70b", + "llama3.1-405b", "gpt-3.5-turbo", "phind-codellama-34b", "gemini-pro", @@ -598,6 +602,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phind-codellama-34b", "wizardcoder-7b", "wizardcoder-13b", @@ -748,6 +755,9 @@ "llama2-13b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "codellama-7b", "codellama-13b", "codellama-34b", @@ -812,6 +822,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phind-codellama-34b", "wizardcoder-7b", "wizardcoder-13b", @@ -860,6 +873,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phi-2", "phind-codellama-34b", "wizardcoder-7b", @@ -947,6 +963,9 @@ "gemma", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "AUTODETECT" ] } @@ -1113,6 +1132,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phind-codellama-34b", "wizardcoder-7b", "wizardcoder-13b", diff --git a/extensions/intellij/src/main/resources/config_schema.json b/extensions/intellij/src/main/resources/config_schema.json index ac7461d8fd..0512f8ac0d 100644 --- a/extensions/intellij/src/main/resources/config_schema.json +++ b/extensions/intellij/src/main/resources/config_schema.json @@ -513,6 +513,9 @@ "anyOf": [ { "enum": [ + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "llama3-8b", "llama3-70b", "codellama-7b", @@ -546,7 +549,8 @@ "enum": [ "gpt-4o", "codestral-latest", - "llama3-70b", + "llama3.1-70b", + "llama3.1-405b", "gpt-3.5-turbo", "phind-codellama-34b", "gemini-pro", @@ -598,6 +602,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phind-codellama-34b", "wizardcoder-7b", "wizardcoder-13b", @@ -748,6 +755,9 @@ "llama2-13b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "codellama-7b", "codellama-13b", "codellama-34b", @@ -812,6 +822,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phind-codellama-34b", "wizardcoder-7b", "wizardcoder-13b", @@ -860,6 +873,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phi-2", "phind-codellama-34b", "wizardcoder-7b", @@ -947,6 +963,9 @@ "gemma", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "AUTODETECT" ] } @@ -1113,6 +1132,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phind-codellama-34b", "wizardcoder-7b", "wizardcoder-13b", diff --git a/extensions/vscode/config_schema.json b/extensions/vscode/config_schema.json index ac7461d8fd..0512f8ac0d 100644 --- a/extensions/vscode/config_schema.json +++ b/extensions/vscode/config_schema.json @@ -513,6 +513,9 @@ "anyOf": [ { "enum": [ + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "llama3-8b", "llama3-70b", "codellama-7b", @@ -546,7 +549,8 @@ "enum": [ "gpt-4o", "codestral-latest", - "llama3-70b", + "llama3.1-70b", + "llama3.1-405b", "gpt-3.5-turbo", "phind-codellama-34b", "gemini-pro", @@ -598,6 +602,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phind-codellama-34b", "wizardcoder-7b", "wizardcoder-13b", @@ -748,6 +755,9 @@ "llama2-13b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "codellama-7b", "codellama-13b", "codellama-34b", @@ -812,6 +822,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phind-codellama-34b", "wizardcoder-7b", "wizardcoder-13b", @@ -860,6 +873,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phi-2", "phind-codellama-34b", "wizardcoder-7b", @@ -947,6 +963,9 @@ "gemma", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "AUTODETECT" ] } @@ -1113,6 +1132,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phind-codellama-34b", "wizardcoder-7b", "wizardcoder-13b", diff --git a/extensions/vscode/continue_rc_schema.json b/extensions/vscode/continue_rc_schema.json index 2c67ce0c19..8e0e499198 100644 --- a/extensions/vscode/continue_rc_schema.json +++ b/extensions/vscode/continue_rc_schema.json @@ -552,6 +552,9 @@ "anyOf": [ { "enum": [ + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "llama3-8b", "llama3-70b", "codellama-7b", @@ -589,7 +592,8 @@ "enum": [ "gpt-4o", "codestral-latest", - "llama3-70b", + "llama3.1-70b", + "llama3.1-405b", "gpt-3.5-turbo", "phind-codellama-34b", "gemini-pro", @@ -645,6 +649,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phind-codellama-34b", "wizardcoder-7b", "wizardcoder-13b", @@ -818,6 +825,9 @@ "llama2-13b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "codellama-7b", "codellama-13b", "codellama-34b", @@ -888,6 +898,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phind-codellama-34b", "wizardcoder-7b", "wizardcoder-13b", @@ -940,6 +953,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phi-2", "phind-codellama-34b", "wizardcoder-7b", @@ -1042,6 +1058,9 @@ "gemma", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "AUTODETECT" ] } @@ -1240,6 +1259,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phind-codellama-34b", "wizardcoder-7b", "wizardcoder-13b", diff --git a/extensions/vscode/package-lock.json b/extensions/vscode/package-lock.json index 118fd459de..e4cf812f76 100644 --- a/extensions/vscode/package-lock.json +++ b/extensions/vscode/package-lock.json @@ -1,12 +1,12 @@ { "name": "continue", - "version": "0.9.189", + "version": "0.9.187", "lockfileVersion": 3, "requires": true, "packages": { "": { "name": "continue", - "version": "0.9.189", + "version": "0.9.187", "license": "Apache-2.0", "dependencies": { "@electron/rebuild": "^3.2.10", diff --git a/gui/src/pages/AddNewModel/AddNewModel.tsx b/gui/src/pages/AddNewModel/AddNewModel.tsx index e08cec157f..ddf947fc26 100644 --- a/gui/src/pages/AddNewModel/AddNewModel.tsx +++ b/gui/src/pages/AddNewModel/AddNewModel.tsx @@ -62,7 +62,7 @@ function AddNewModel() { const [providersSelected, setProvidersSelected] = React.useState(true); return ( -
+
Add a new model

- - To add a new model you can either: -
    -
  • - Start by selecting and configuring a provider, then choosing your - model -
  • -
  • Select a specific model directly
  • -
- - Visit our setup overview docs - {" "} - to learn more. -
- -
- { - setProvidersSelected((prev) => !prev); - }} - > -
+
+ + To add a new model you can either: +
    +
  • + Start by selecting and configuring a provider, then choosing your + model +
  • +
  • Select a specific model directly
  • +
+ + Visit our setup overview docs + {" "} + to learn more. +
- {providersSelected ? ( - <> -
-

Providers

-

- Select a provider below, or configure your own in{" "} - config.json -

-
+
+ { + setProvidersSelected((prev) => !prev); + }} + > +
- - {Object.entries(providers).map(([providerName, modelInfo], i) => ( - { - console.log(`/addModel/provider/${providerName}`); - navigate(`/addModel/provider/${providerName}`); - }} - /> - ))} - - - ) : ( - <> -
-

Models

-

- Select a model from the most popular options below, or configure - your own in config.json -

-
+ {providersSelected ? ( + <> +
+

Providers

+

+ Select a provider below, or configure your own in{" "} + config.json +

+
- {Object.entries(modelsByProvider).map( - ([providerTitle, modelConfigsByProviderTitle]) => ( -
-
-

{providerTitle}

-
-
+ + {Object.entries(providers).map(([providerName, modelInfo], i) => ( + { + console.log(`/addModel/provider/${providerName}`); + navigate(`/addModel/provider/${providerName}`); + }} + /> + ))} + + + ) : ( + <> +
+

Models

+

+ Select a model from the most popular options below, or configure + your own in config.json +

+
- - {modelConfigsByProviderTitle.map((config) => ( - { - const model = { - ...config.params, - ..._.merge( - {}, - ...(config.dimensions?.map((dimension, i) => { - if (!dimensionChoices?.[i]) return {}; - return { - ...dimension.options[dimensionChoices[i]], - }; - }) || []), - ), - provider: providers[selectedProvider].provider, - }; - ideMessenger.post("config/addModel", { model }); - dispatch( - setDefaultModel({ - title: model.title, - force: true, - }), - ); - navigate("/"); + {Object.entries(modelsByProvider).map( + ([providerTitle, modelConfigsByProviderTitle]) => ( +
+
+

{providerTitle}

+
- ))} - -
- ), - )} - - )} +
- { - ideMessenger.post("openConfigJson", undefined); - }} - > -

- - Open config.json -

-
+ + {modelConfigsByProviderTitle.map((config) => ( + { + const model = { + ...config.params, + ..._.merge( + {}, + ...(config.dimensions?.map((dimension, i) => { + if (!dimensionChoices?.[i]) return {}; + return { + ...dimension.options[dimensionChoices[i]], + }; + }) || []), + ), + provider: providers[selectedProvider].provider, + }; + ideMessenger.post("config/addModel", { model }); + dispatch( + setDefaultModel({ + title: model.title, + force: true, + }), + ); + navigate("/"); + }} + /> + ))} + +
+ ), + )} + + )} + + { + ideMessenger.post("openConfigJson", undefined); + }} + > +

+ + Open config.json +

+
+
); } diff --git a/gui/src/pages/AddNewModel/configs/models.ts b/gui/src/pages/AddNewModel/configs/models.ts index 62fa6bb9ec..237e69b83b 100644 --- a/gui/src/pages/AddNewModel/configs/models.ts +++ b/gui/src/pages/AddNewModel/configs/models.ts @@ -32,6 +32,113 @@ export interface ModelPackage { } export const models: { [key: string]: ModelPackage } = { + llama31Chat: { + title: "Llama3.1 Chat", + description: "The latest model from Meta, fine-tuned for chat", + refUrl: "", + params: { + title: "Llama3.1-8b", + model: "llama3.1-8b", + contextLength: 8192, + }, + icon: "meta.png", + dimensions: [ + { + name: "Parameter Count", + description: "The number of parameters in the model", + options: { + "8b": { + model: "llama3.1-8b", + title: "Llama3.1-8b", + }, + "70b": { + model: "llama3.1-70b", + title: "Llama3.1-70b", + }, + "405b": { + model: "llama3.1-405b", + title: "Llama3.1-405b", + }, + }, + }, + ], + providerOptions: [ + "ollama", + "lmstudio", + "together", + "llama.cpp", + "replicate", + ], + isOpenSource: true, + }, + deepseek: { + title: "DeepSeek-Coder", + description: + "A model pre-trained on 2 trillion tokens including 80+ programming languages and a repo-level corpus.", + params: { + title: "DeepSeek-7b", + model: "deepseek-7b", + contextLength: 4096, + }, + icon: "deepseek.png", + dimensions: [ + { + name: "Parameter Count", + description: "The number of parameters in the model", + options: { + "1b": { + model: "deepseek-1b", + title: "DeepSeek-1b", + }, + "7b": { + model: "deepseek-7b", + title: "DeepSeek-7b", + }, + "33b": { + model: "deepseek-33b", + title: "DeepSeek-33b", + }, + }, + }, + ], + providerOptions: ["ollama", "lmstudio", "llama.cpp"], + isOpenSource: true, + }, + mistralOs: { + title: "Mistral", + description: + "A series of open-weight models created by Mistral AI, highly competent for code generation and other tasks", + params: { + title: "Mistral", + model: "mistral-7b", + contextLength: 4096, + }, + dimensions: [ + { + name: "Parameter Count", + description: "The number of parameters in the model", + options: { + "7b": { + model: "mistral-7b", + title: "Mistral-7b", + }, + "8x7b (MoE)": { + model: "mistral-8x7b", + title: "Mixtral", + }, + }, + }, + ], + icon: "mistral.png", + providerOptions: [ + "ollama", + "lmstudio", + "together", + "llama.cpp", + "replicate", + ], + isOpenSource: true, + }, codeLlamaInstruct: { title: "CodeLlama Instruct", description: @@ -76,7 +183,7 @@ export const models: { [key: string]: ModelPackage } = { ], isOpenSource: true, }, - codellama70bTrial: { + llama3170bTrial: { title: "Codellama 70b (Free Trial)", description: "The best code model from Meta, fine-tuned for code generation and conversation", @@ -90,6 +197,19 @@ export const models: { [key: string]: ModelPackage } = { providerOptions: ["free-trial"], isOpenSource: false, }, + llama31405bTrial: { + title: "Llama3.1 405b (Free Trial)", + description: "The latest Llama model from Meta, fine-tuned for chat", + refUrl: "", + params: { + title: "Llama3.1-405b", + model: "llama3.1-405b", + contextLength: 8192, + }, + icon: "meta.png", + providerOptions: ["free-trial"], + isOpenSource: false, + }, mixtralTrial: { title: "Mixtral (Free Trial)", description: @@ -130,6 +250,45 @@ export const models: { [key: string]: ModelPackage } = { providerOptions: ["groq"], isOpenSource: false, }, + llama318bChat: { + title: "Llama3.1 8b Chat", + description: "The latest Llama model from Meta, fine-tuned for chat", + refUrl: "", + params: { + title: "Llama3.1-8b", + model: "llama3.1-8b", + contextLength: 8192, + }, + icon: "meta.png", + providerOptions: ["groq"], + isOpenSource: false, + }, + llama3170bChat: { + title: "Llama3.1 70b Chat", + description: "The latest Llama model from Meta, fine-tuned for chat", + refUrl: "", + params: { + title: "Llama3.1-70b", + model: "llama3.1-70b", + contextLength: 8192, + }, + icon: "meta.png", + providerOptions: ["groq"], + isOpenSource: false, + }, + llama31405bChat: { + title: "Llama3.1 405b Chat", + description: "The latest Llama model from Meta, fine-tuned for chat", + refUrl: "", + params: { + title: "Llama3.1-405b", + model: "llama3.1-405b", + contextLength: 8192, + }, + icon: "meta.png", + providerOptions: ["groq"], + isOpenSource: false, + }, llama270bChat: { title: "Llama2 70b Chat", description: "The latest Llama model from Meta, fine-tuned for chat", @@ -230,41 +389,6 @@ export const models: { [key: string]: ModelPackage } = { ], isOpenSource: true, }, - mistralOs: { - title: "Mistral", - description: - "A series of open-weight models created by Mistral AI, highly competent for code generation and other tasks", - params: { - title: "Mistral", - model: "mistral-7b", - contextLength: 4096, - }, - dimensions: [ - { - name: "Parameter Count", - description: "The number of parameters in the model", - options: { - "7b": { - model: "mistral-7b", - title: "Mistral-7b", - }, - "8x7b (MoE)": { - model: "mistral-8x7b", - title: "Mixtral", - }, - }, - }, - ], - icon: "mistral.png", - providerOptions: [ - "ollama", - "lmstudio", - "together", - "llama.cpp", - "replicate", - ], - isOpenSource: true, - }, codestral: { title: "Codestral", description: @@ -395,39 +519,6 @@ export const models: { [key: string]: ModelPackage } = { providerOptions: ["gemini"], isOpenSource: false, }, - deepseek: { - title: "DeepSeek-Coder", - description: - "A model pre-trained on 2 trillion tokens including 80+ programming languages and a repo-level corpus.", - params: { - title: "DeepSeek-7b", - model: "deepseek-7b", - contextLength: 4096, - }, - icon: "deepseek.png", - dimensions: [ - { - name: "Parameter Count", - description: "The number of parameters in the model", - options: { - "1b": { - model: "deepseek-1b", - title: "DeepSeek-1b", - }, - "7b": { - model: "deepseek-7b", - title: "DeepSeek-7b", - }, - "33b": { - model: "deepseek-33b", - title: "DeepSeek-33b", - }, - }, - }, - ], - providerOptions: ["ollama", "lmstudio", "llama.cpp"], - isOpenSource: true, - }, commandR: { title: "Command R", description: diff --git a/gui/src/pages/AddNewModel/configs/providers.ts b/gui/src/pages/AddNewModel/configs/providers.ts index 975170f98b..630b5de786 100644 --- a/gui/src/pages/AddNewModel/configs/providers.ts +++ b/gui/src/pages/AddNewModel/configs/providers.ts @@ -262,8 +262,9 @@ Select the \`GPT-4o\` model below to complete your provider configuration, but n }, ], packages: [ - models.llama370bChat, - models.llama38bChat, + models.llama31405bChat, + models.llama3170bChat, + models.llama318bChat, { ...models.mixtralTrial, title: "Mixtral" }, models.llama270bChat, { @@ -300,7 +301,7 @@ Select the \`GPT-4o\` model below to complete your provider configuration, but n ...completionParamsInputsConfigs, ], packages: [ - models.llama3Chat, + models.llama31Chat, models.codeLlamaInstruct, models.mistralOs, ].map((p) => { @@ -464,7 +465,8 @@ After it's up and running, you can start using Continue.`, icon: "openai.png", tags: [ModelProviderTags.Free], packages: [ - models.codellama70bTrial, + models.llama31405bTrial, + models.llama3170bTrial, { ...models.claude35Sonnet, title: "Claude 3.5 Sonnet (trial)" }, { ...models.gpt4o, title: "GPT-4o (trial)" }, { ...models.gpt35turbo, title: "GPT-3.5-Turbo (trial)" },