diff --git a/frontend/src/components/LLMSelection/OpenAiOptions/index.jsx b/frontend/src/components/LLMSelection/OpenAiOptions/index.jsx index cbd83edb..ffc0f91e 100644 --- a/frontend/src/components/LLMSelection/OpenAiOptions/index.jsx +++ b/frontend/src/components/LLMSelection/OpenAiOptions/index.jsx @@ -81,19 +81,23 @@ function OpenAIModelSelection({ apiKey, settings }) { className="bg-zinc-900 border border-gray-500 text-white text-sm rounded-lg block w-full p-2.5" > - {["gpt-3.5-turbo", "gpt-4", "gpt-4-1106-preview", "gpt-4-32k"].map( - (model) => { - return ( - - ); - } - )} + {[ + "gpt-3.5-turbo", + "gpt-3.5-turbo-1106", + "gpt-4", + "gpt-4-1106-preview", + "gpt-4-32k", + ].map((model) => { + return ( + + ); + })} {customModels.length > 0 && ( diff --git a/frontend/src/components/Modals/MangeWorkspace/Settings/ChatModelPreference/useGetProviderModels.js b/frontend/src/components/Modals/MangeWorkspace/Settings/ChatModelPreference/useGetProviderModels.js index eae1b4ad..270f8a96 100644 --- a/frontend/src/components/Modals/MangeWorkspace/Settings/ChatModelPreference/useGetProviderModels.js +++ b/frontend/src/components/Modals/MangeWorkspace/Settings/ChatModelPreference/useGetProviderModels.js @@ -4,7 +4,13 @@ import { useEffect, useState } from "react"; // Providers which cannot use this feature for workspace<>model selection export const DISABLED_PROVIDERS = ["azure", "lmstudio"]; const PROVIDER_DEFAULT_MODELS = { - openai: ["gpt-3.5-turbo", "gpt-4", "gpt-4-1106-preview", "gpt-4-32k"], + openai: [ + "gpt-3.5-turbo", + "gpt-3.5-turbo-1106", + "gpt-4", + "gpt-4-1106-preview", + "gpt-4-32k", + ], gemini: ["gemini-pro"], anthropic: ["claude-2", "claude-instant-1"], azure: [], diff --git a/server/utils/AiProviders/openAi/index.js b/server/utils/AiProviders/openAi/index.js index 582bc054..a25f69ea 100644 --- a/server/utils/AiProviders/openAi/index.js +++ b/server/utils/AiProviders/openAi/index.js @@ -46,6 +46,8 @@ class OpenAiLLM { switch (this.model) { case "gpt-3.5-turbo": return 4096; + case "gpt-3.5-turbo-1106": + return 16385; case "gpt-4": return 8192; case "gpt-4-1106-preview": @@ -61,6 +63,7 @@ class OpenAiLLM { const validModels = [ "gpt-4", "gpt-3.5-turbo", + "gpt-3.5-turbo-1106", "gpt-4-1106-preview", "gpt-4-32k", ];