diff --git a/core/autocomplete/templates.ts b/core/autocomplete/templates.ts index 4de43739ab..c0177bb12f 100644 --- a/core/autocomplete/templates.ts +++ b/core/autocomplete/templates.ts @@ -155,11 +155,6 @@ const deepseekFimTemplate: AutocompleteTemplate = { }, }; -const deepseekFimTemplateWrongPipeChar: AutocompleteTemplate = { - template: "<|fim▁begin|>{{{prefix}}}<|fim▁hole|>{{{suffix}}}<|fim▁end|>", - completionOptions: { stop: ["<|fim▁begin|>", "<|fim▁hole|>", "<|fim▁end|>"] }, -}; - const gptAutocompleteTemplate: AutocompleteTemplate = { template: `\`\`\` {{{prefix}}}[BLANK]{{{suffix}}} diff --git a/core/llm/llms/FreeTrial.ts b/core/llm/llms/FreeTrial.ts index b2db6a5a17..e4c1674e63 100644 --- a/core/llm/llms/FreeTrial.ts +++ b/core/llm/llms/FreeTrial.ts @@ -165,7 +165,8 @@ class FreeTrial extends BaseLLM { return [ "codestral-latest", "claude-3-5-sonnet-20240620", - "llama3-70b", + "llama3.1-405b", + "llama3.1-70b", "gpt-4o", "gpt-3.5-turbo", "claude-3-haiku-20240307", diff --git a/core/llm/llms/Groq.ts b/core/llm/llms/Groq.ts index 0aa5b56365..27c51e1e33 100644 --- a/core/llm/llms/Groq.ts +++ b/core/llm/llms/Groq.ts @@ -14,6 +14,9 @@ class Groq extends OpenAI { gemma: "gemma-7b-it", "llama3-8b": "llama3-8b-8192", "llama3-70b": "llama3-70b-8192", + "llama3.1-8b": "llama-3.1-8b-instant", + "llama3.1-70b": "llama-3.1-70b-versatile", + "llama3.1-405b": "llama-3.1-405b-reasoning", }; protected _convertModelName(model: string): string { return Groq.modelConversion[model] ?? model; diff --git a/core/llm/llms/Ollama.ts b/core/llm/llms/Ollama.ts index e24a489f0c..8da67a14e9 100644 --- a/core/llm/llms/Ollama.ts +++ b/core/llm/llms/Ollama.ts @@ -88,6 +88,9 @@ class Ollama extends BaseLLM { "codellama-70b": "codellama:70b", "llama3-8b": "llama3:8b", "llama3-70b": "llama3:70b", + "llama3.1-8b": "llama3.1:8b", + "llama3.1-70b": "llama3.1:70b", + "llama3.1-405b": "llama3.1:405b", "phi-2": "phi:2.7b", "phind-codellama-34b": "phind-codellama:34b-v2", "wizardcoder-7b": "wizardcoder:7b-python", diff --git a/core/llm/llms/Replicate.ts b/core/llm/llms/Replicate.ts index 7cfa6e8901..5dde240148 100644 --- a/core/llm/llms/Replicate.ts +++ b/core/llm/llms/Replicate.ts @@ -18,6 +18,9 @@ class Replicate extends BaseLLM { "llama2-13b": "meta/llama-2-13b-chat" as any, "llama3-8b": "meta/meta-llama-3-8b-instruct" as any, "llama3-70b": "meta/meta-llama-3-70b-instruct" as any, + "llama3.1-8b": "meta/meta-llama-3.1-8b-instruct" as any, + "llama3.1-70b": "meta/meta-llama-3.1-70b-instruct" as any, + "llama3.1-405b": "meta/meta-llama-3.1-405b-instruct" as any, "zephyr-7b": "nateraw/zephyr-7b-beta:b79f33de5c6c4e34087d44eaea4a9d98ce5d3f3a09522f7328eea0685003a931", "mistral-7b": diff --git a/core/llm/llms/Together.ts b/core/llm/llms/Together.ts index 277389c770..eaa1442771 100644 --- a/core/llm/llms/Together.ts +++ b/core/llm/llms/Together.ts @@ -19,6 +19,9 @@ class Together extends OpenAI { "codellama-70b": "codellama/CodeLlama-70b-Instruct-hf", "llama3-8b": "meta-llama/Llama-3-8b-chat-hf", "llama3-70b": "meta-llama/Llama-3-70b-chat-hf", + "llama3.1-8b": "meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo", + "llama3.1-70b": "meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo", + "llama3.1-405b": "meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo", "llama2-7b": "togethercomputer/llama-2-7b-chat", "llama2-13b": "togethercomputer/llama-2-13b-chat", "llama2-70b": "togethercomputer/llama-2-70b-chat", diff --git a/docs/static/schemas/config.json b/docs/static/schemas/config.json index ac7461d8fd..0512f8ac0d 100644 --- a/docs/static/schemas/config.json +++ b/docs/static/schemas/config.json @@ -513,6 +513,9 @@ "anyOf": [ { "enum": [ + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "llama3-8b", "llama3-70b", "codellama-7b", @@ -546,7 +549,8 @@ "enum": [ "gpt-4o", "codestral-latest", - "llama3-70b", + "llama3.1-70b", + "llama3.1-405b", "gpt-3.5-turbo", "phind-codellama-34b", "gemini-pro", @@ -598,6 +602,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phind-codellama-34b", "wizardcoder-7b", "wizardcoder-13b", @@ -748,6 +755,9 @@ "llama2-13b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "codellama-7b", "codellama-13b", "codellama-34b", @@ -812,6 +822,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phind-codellama-34b", "wizardcoder-7b", "wizardcoder-13b", @@ -860,6 +873,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phi-2", "phind-codellama-34b", "wizardcoder-7b", @@ -947,6 +963,9 @@ "gemma", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "AUTODETECT" ] } @@ -1113,6 +1132,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phind-codellama-34b", "wizardcoder-7b", "wizardcoder-13b", diff --git a/extensions/intellij/src/main/resources/config_schema.json b/extensions/intellij/src/main/resources/config_schema.json index ac7461d8fd..0512f8ac0d 100644 --- a/extensions/intellij/src/main/resources/config_schema.json +++ b/extensions/intellij/src/main/resources/config_schema.json @@ -513,6 +513,9 @@ "anyOf": [ { "enum": [ + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "llama3-8b", "llama3-70b", "codellama-7b", @@ -546,7 +549,8 @@ "enum": [ "gpt-4o", "codestral-latest", - "llama3-70b", + "llama3.1-70b", + "llama3.1-405b", "gpt-3.5-turbo", "phind-codellama-34b", "gemini-pro", @@ -598,6 +602,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phind-codellama-34b", "wizardcoder-7b", "wizardcoder-13b", @@ -748,6 +755,9 @@ "llama2-13b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "codellama-7b", "codellama-13b", "codellama-34b", @@ -812,6 +822,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phind-codellama-34b", "wizardcoder-7b", "wizardcoder-13b", @@ -860,6 +873,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phi-2", "phind-codellama-34b", "wizardcoder-7b", @@ -947,6 +963,9 @@ "gemma", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "AUTODETECT" ] } @@ -1113,6 +1132,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phind-codellama-34b", "wizardcoder-7b", "wizardcoder-13b", diff --git a/extensions/vscode/config_schema.json b/extensions/vscode/config_schema.json index ac7461d8fd..0512f8ac0d 100644 --- a/extensions/vscode/config_schema.json +++ b/extensions/vscode/config_schema.json @@ -513,6 +513,9 @@ "anyOf": [ { "enum": [ + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "llama3-8b", "llama3-70b", "codellama-7b", @@ -546,7 +549,8 @@ "enum": [ "gpt-4o", "codestral-latest", - "llama3-70b", + "llama3.1-70b", + "llama3.1-405b", "gpt-3.5-turbo", "phind-codellama-34b", "gemini-pro", @@ -598,6 +602,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phind-codellama-34b", "wizardcoder-7b", "wizardcoder-13b", @@ -748,6 +755,9 @@ "llama2-13b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "codellama-7b", "codellama-13b", "codellama-34b", @@ -812,6 +822,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phind-codellama-34b", "wizardcoder-7b", "wizardcoder-13b", @@ -860,6 +873,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phi-2", "phind-codellama-34b", "wizardcoder-7b", @@ -947,6 +963,9 @@ "gemma", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "AUTODETECT" ] } @@ -1113,6 +1132,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phind-codellama-34b", "wizardcoder-7b", "wizardcoder-13b", diff --git a/extensions/vscode/continue_rc_schema.json b/extensions/vscode/continue_rc_schema.json index 2c67ce0c19..8e0e499198 100644 --- a/extensions/vscode/continue_rc_schema.json +++ b/extensions/vscode/continue_rc_schema.json @@ -552,6 +552,9 @@ "anyOf": [ { "enum": [ + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "llama3-8b", "llama3-70b", "codellama-7b", @@ -589,7 +592,8 @@ "enum": [ "gpt-4o", "codestral-latest", - "llama3-70b", + "llama3.1-70b", + "llama3.1-405b", "gpt-3.5-turbo", "phind-codellama-34b", "gemini-pro", @@ -645,6 +649,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phind-codellama-34b", "wizardcoder-7b", "wizardcoder-13b", @@ -818,6 +825,9 @@ "llama2-13b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "codellama-7b", "codellama-13b", "codellama-34b", @@ -888,6 +898,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phind-codellama-34b", "wizardcoder-7b", "wizardcoder-13b", @@ -940,6 +953,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phi-2", "phind-codellama-34b", "wizardcoder-7b", @@ -1042,6 +1058,9 @@ "gemma", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "AUTODETECT" ] } @@ -1240,6 +1259,9 @@ "codellama-70b", "llama3-8b", "llama3-70b", + "llama3.1-8b", + "llama3.1-70b", + "llama3.1-405b", "phind-codellama-34b", "wizardcoder-7b", "wizardcoder-13b", diff --git a/extensions/vscode/package-lock.json b/extensions/vscode/package-lock.json index 118fd459de..e4cf812f76 100644 --- a/extensions/vscode/package-lock.json +++ b/extensions/vscode/package-lock.json @@ -1,12 +1,12 @@ { "name": "continue", - "version": "0.9.189", + "version": "0.9.187", "lockfileVersion": 3, "requires": true, "packages": { "": { "name": "continue", - "version": "0.9.189", + "version": "0.9.187", "license": "Apache-2.0", "dependencies": { "@electron/rebuild": "^3.2.10", diff --git a/gui/src/pages/AddNewModel/AddNewModel.tsx b/gui/src/pages/AddNewModel/AddNewModel.tsx index e08cec157f..ddf947fc26 100644 --- a/gui/src/pages/AddNewModel/AddNewModel.tsx +++ b/gui/src/pages/AddNewModel/AddNewModel.tsx @@ -62,7 +62,7 @@ function AddNewModel() { const [providersSelected, setProvidersSelected] = React.useState(true); return ( -
- Select a provider below, or configure your own in{" "}
- config.json
-
- Select a model from the most popular options below, or configure
- your own in config.json
-
+ Select a provider below, or configure your own in{" "}
+ config.json
+
+ Select a model from the most popular options below, or configure
+ your own in config.json
+