function getVectorDbClass() { const vectorSelection = process.env.VECTOR_DB || "pinecone"; switch (vectorSelection) { case "pinecone": const { Pinecone } = require("../vectorDbProviders/pinecone"); return Pinecone; case "chroma": const { Chroma } = require("../vectorDbProviders/chroma"); return Chroma; case "lancedb": const { LanceDb } = require("../vectorDbProviders/lance"); return LanceDb; case "weaviate": const { Weaviate } = require("../vectorDbProviders/weaviate"); return Weaviate; case "qdrant": const { QDrant } = require("../vectorDbProviders/qdrant"); return QDrant; case "milvus": const { Milvus } = require("../vectorDbProviders/milvus"); return Milvus; case "zilliz": const { Zilliz } = require("../vectorDbProviders/zilliz"); return Zilliz; case "astra": const { AstraDB } = require("../vectorDbProviders/astra"); return AstraDB; default: throw new Error("ENV: No VECTOR_DB value found in environment!"); } } function getLLMProvider(modelPreference = null) { const vectorSelection = process.env.LLM_PROVIDER || "openai"; const embedder = getEmbeddingEngineSelection(); switch (vectorSelection) { case "openai": const { OpenAiLLM } = require("../AiProviders/openAi"); return new OpenAiLLM(embedder, modelPreference); case "azure": const { AzureOpenAiLLM } = require("../AiProviders/azureOpenAi"); return new AzureOpenAiLLM(embedder, modelPreference); case "anthropic": const { AnthropicLLM } = require("../AiProviders/anthropic"); return new AnthropicLLM(embedder, modelPreference); case "gemini": const { GeminiLLM } = require("../AiProviders/gemini"); return new GeminiLLM(embedder, modelPreference); case "lmstudio": const { LMStudioLLM } = require("../AiProviders/lmStudio"); return new LMStudioLLM(embedder, modelPreference); case "localai": const { LocalAiLLM } = require("../AiProviders/localAi"); return new LocalAiLLM(embedder, modelPreference); case "ollama": const { OllamaAILLM } = require("../AiProviders/ollama"); return new OllamaAILLM(embedder, modelPreference); case "togetherai": const { TogetherAiLLM } = require("../AiProviders/togetherAi"); return new TogetherAiLLM(embedder, modelPreference); case "mistral": const { MistralLLM } = require("../AiProviders/mistral"); return new MistralLLM(embedder, modelPreference); case "native": const { NativeLLM } = require("../AiProviders/native"); return new NativeLLM(embedder, modelPreference); case "huggingface": const { HuggingFaceLLM } = require("../AiProviders/huggingface"); return new HuggingFaceLLM(embedder, modelPreference); default: throw new Error("ENV: No LLM_PROVIDER value found in environment!"); } } function getEmbeddingEngineSelection() { const engineSelection = process.env.EMBEDDING_ENGINE; switch (engineSelection) { case "openai": const { OpenAiEmbedder } = require("../EmbeddingEngines/openAi"); return new OpenAiEmbedder(); case "azure": const { AzureOpenAiEmbedder, } = require("../EmbeddingEngines/azureOpenAi"); return new AzureOpenAiEmbedder(); case "localai": const { LocalAiEmbedder } = require("../EmbeddingEngines/localAi"); return new LocalAiEmbedder(); case "native": const { NativeEmbedder } = require("../EmbeddingEngines/native"); console.log("\x1b[34m[INFO]\x1b[0m Using Native Embedder"); return new NativeEmbedder(); default: return null; } } // Some models have lower restrictions on chars that can be encoded in a single pass // and by default we assume it can handle 1,000 chars, but some models use work with smaller // chars so here we can override that value when embedding information. function maximumChunkLength() { if ( !!process.env.EMBEDDING_MODEL_MAX_CHUNK_LENGTH && !isNaN(process.env.EMBEDDING_MODEL_MAX_CHUNK_LENGTH) && Number(process.env.EMBEDDING_MODEL_MAX_CHUNK_LENGTH) > 1 ) return Number(process.env.EMBEDDING_MODEL_MAX_CHUNK_LENGTH); return 1_000; } function toChunks(arr, size) { return Array.from({ length: Math.ceil(arr.length / size) }, (_v, i) => arr.slice(i * size, i * size + size) ); } module.exports = { getEmbeddingEngineSelection, maximumChunkLength, getVectorDbClass, getLLMProvider, toChunks, };