mirror of
https://github.com/Mintplex-Labs/anything-llm.git
synced 2024-11-10 17:00:11 +01:00
dc4ad6b5a9
* wip bg workers for live document sync * Add ability to re-embed specific documents across many workspaces via background queue bgworkser is gated behind expieremental system setting flag that needs to be explictly enabled UI for watching/unwatching docments that are embedded. TODO: UI to easily manage all bg tasks and see run results TODO: UI to enable this feature and background endpoints to manage it * create frontend views and paths Move elements to correct experimental scope * update migration to delete runs on removal of watched document * Add watch support to YouTube transcripts (#1716) * Add watch support to YouTube transcripts refactor how sync is done for supported types * Watch specific files in Confluence space (#1718) Add failure-prune check for runs * create tmp workflow modifications for beta image * create tmp workflow modifications for beta image * create tmp workflow modifications for beta image * dual build update copy of alert modals * update job interval * Add support for live-sync of Github files * update copy for document sync feature * hide Experimental features from UI * update docs links * [FEAT] Implement new settings menu for experimental features (#1735) * implement new settings menu for experimental features * remove unused context save bar --------- Co-authored-by: timothycarambat <rambat1010@gmail.com> * dont run job on boot * unset workflow changes * Add persistent encryption service Relay key to collector so persistent encryption can be used Encrypt any private data in chunkSources used for replay during resync jobs * update jsDOC * Linting and organization * update modal copy for feature --------- Co-authored-by: Sean Hatfield <seanhatfield5@gmail.com>
252 lines
8.1 KiB
Plaintext
252 lines
8.1 KiB
Plaintext
SERVER_PORT=3001
|
|
STORAGE_DIR="/app/server/storage"
|
|
UID='1000'
|
|
GID='1000'
|
|
# SIG_KEY='passphrase' # Please generate random string at least 32 chars long.
|
|
# SIG_SALT='salt' # Please generate random string at least 32 chars long.
|
|
# JWT_SECRET="my-random-string-for-seeding" # Only needed if AUTH_TOKEN is set. Please generate random string at least 12 chars long.
|
|
|
|
###########################################
|
|
######## LLM API SElECTION ################
|
|
###########################################
|
|
# LLM_PROVIDER='openai'
|
|
# OPEN_AI_KEY=
|
|
# OPEN_MODEL_PREF='gpt-4o'
|
|
|
|
# LLM_PROVIDER='gemini'
|
|
# GEMINI_API_KEY=
|
|
# GEMINI_LLM_MODEL_PREF='gemini-pro'
|
|
|
|
# LLM_PROVIDER='azure'
|
|
# AZURE_OPENAI_ENDPOINT=
|
|
# AZURE_OPENAI_KEY=
|
|
# OPEN_MODEL_PREF='my-gpt35-deployment' # This is the "deployment" on Azure you want to use. Not the base model.
|
|
# EMBEDDING_MODEL_PREF='embedder-model' # This is the "deployment" on Azure you want to use for embeddings. Not the base model. Valid base model is text-embedding-ada-002
|
|
|
|
# LLM_PROVIDER='anthropic'
|
|
# ANTHROPIC_API_KEY=sk-ant-xxxx
|
|
# ANTHROPIC_MODEL_PREF='claude-2'
|
|
|
|
# LLM_PROVIDER='lmstudio'
|
|
# LMSTUDIO_BASE_PATH='http://your-server:1234/v1'
|
|
# LMSTUDIO_MODEL_PREF='Loaded from Chat UI' # this is a bug in LMStudio 0.2.17
|
|
# LMSTUDIO_MODEL_TOKEN_LIMIT=4096
|
|
|
|
# LLM_PROVIDER='localai'
|
|
# LOCAL_AI_BASE_PATH='http://host.docker.internal:8080/v1'
|
|
# LOCAL_AI_MODEL_PREF='luna-ai-llama2'
|
|
# LOCAL_AI_MODEL_TOKEN_LIMIT=4096
|
|
# LOCAL_AI_API_KEY="sk-123abc"
|
|
|
|
# LLM_PROVIDER='ollama'
|
|
# OLLAMA_BASE_PATH='http://host.docker.internal:11434'
|
|
# OLLAMA_MODEL_PREF='llama2'
|
|
# OLLAMA_MODEL_TOKEN_LIMIT=4096
|
|
|
|
# LLM_PROVIDER='togetherai'
|
|
# TOGETHER_AI_API_KEY='my-together-ai-key'
|
|
# TOGETHER_AI_MODEL_PREF='mistralai/Mixtral-8x7B-Instruct-v0.1'
|
|
|
|
# LLM_PROVIDER='mistral'
|
|
# MISTRAL_API_KEY='example-mistral-ai-api-key'
|
|
# MISTRAL_MODEL_PREF='mistral-tiny'
|
|
|
|
# LLM_PROVIDER='perplexity'
|
|
# PERPLEXITY_API_KEY='my-perplexity-key'
|
|
# PERPLEXITY_MODEL_PREF='codellama-34b-instruct'
|
|
|
|
# LLM_PROVIDER='openrouter'
|
|
# OPENROUTER_API_KEY='my-openrouter-key'
|
|
# OPENROUTER_MODEL_PREF='openrouter/auto'
|
|
|
|
# LLM_PROVIDER='huggingface'
|
|
# HUGGING_FACE_LLM_ENDPOINT=https://uuid-here.us-east-1.aws.endpoints.huggingface.cloud
|
|
# HUGGING_FACE_LLM_API_KEY=hf_xxxxxx
|
|
# HUGGING_FACE_LLM_TOKEN_LIMIT=8000
|
|
|
|
# LLM_PROVIDER='groq'
|
|
# GROQ_API_KEY=gsk_abcxyz
|
|
# GROQ_MODEL_PREF=llama3-8b-8192
|
|
|
|
# LLM_PROVIDER='koboldcpp'
|
|
# KOBOLD_CPP_BASE_PATH='http://127.0.0.1:5000/v1'
|
|
# KOBOLD_CPP_MODEL_PREF='koboldcpp/codellama-7b-instruct.Q4_K_S'
|
|
# KOBOLD_CPP_MODEL_TOKEN_LIMIT=4096
|
|
|
|
# LLM_PROVIDER='textgenwebui'
|
|
# TEXT_GEN_WEB_UI_BASE_PATH='http://127.0.0.1:5000/v1'
|
|
# TEXT_GEN_WEB_UI_TOKEN_LIMIT=4096
|
|
# TEXT_GEN_WEB_UI_API_KEY='sk-123abc'
|
|
|
|
# LLM_PROVIDER='generic-openai'
|
|
# GENERIC_OPEN_AI_BASE_PATH='http://proxy.url.openai.com/v1'
|
|
# GENERIC_OPEN_AI_MODEL_PREF='gpt-3.5-turbo'
|
|
# GENERIC_OPEN_AI_MODEL_TOKEN_LIMIT=4096
|
|
# GENERIC_OPEN_AI_API_KEY=sk-123abc
|
|
|
|
# LLM_PROVIDER='litellm'
|
|
# LITE_LLM_MODEL_PREF='gpt-3.5-turbo'
|
|
# LITE_LLM_MODEL_TOKEN_LIMIT=4096
|
|
# LITE_LLM_BASE_PATH='http://127.0.0.1:4000'
|
|
# LITE_LLM_API_KEY='sk-123abc'
|
|
|
|
# LLM_PROVIDER='cohere'
|
|
# COHERE_API_KEY=
|
|
# COHERE_MODEL_PREF='command-r'
|
|
|
|
###########################################
|
|
######## Embedding API SElECTION ##########
|
|
###########################################
|
|
# Only used if you are using an LLM that does not natively support embedding (openai or Azure)
|
|
# EMBEDDING_ENGINE='openai'
|
|
# OPEN_AI_KEY=sk-xxxx
|
|
# EMBEDDING_MODEL_PREF='text-embedding-ada-002'
|
|
|
|
# EMBEDDING_ENGINE='azure'
|
|
# AZURE_OPENAI_ENDPOINT=
|
|
# AZURE_OPENAI_KEY=
|
|
# EMBEDDING_MODEL_PREF='my-embedder-model' # This is the "deployment" on Azure you want to use for embeddings. Not the base model. Valid base model is text-embedding-ada-002
|
|
|
|
# EMBEDDING_ENGINE='localai'
|
|
# EMBEDDING_BASE_PATH='http://localhost:8080/v1'
|
|
# EMBEDDING_MODEL_PREF='text-embedding-ada-002'
|
|
# EMBEDDING_MODEL_MAX_CHUNK_LENGTH=1000 # The max chunk size in chars a string to embed can be
|
|
|
|
# EMBEDDING_ENGINE='ollama'
|
|
# EMBEDDING_BASE_PATH='http://host.docker.internal:11434'
|
|
# EMBEDDING_MODEL_PREF='nomic-embed-text:latest'
|
|
# EMBEDDING_MODEL_MAX_CHUNK_LENGTH=8192
|
|
|
|
# EMBEDDING_ENGINE='lmstudio'
|
|
# EMBEDDING_BASE_PATH='https://host.docker.internal:1234/v1'
|
|
# EMBEDDING_MODEL_PREF='nomic-ai/nomic-embed-text-v1.5-GGUF/nomic-embed-text-v1.5.Q4_0.gguf'
|
|
# EMBEDDING_MODEL_MAX_CHUNK_LENGTH=8192
|
|
|
|
# EMBEDDING_ENGINE='cohere'
|
|
# COHERE_API_KEY=
|
|
# EMBEDDING_MODEL_PREF='embed-english-v3.0'
|
|
|
|
# EMBEDDING_ENGINE='voyageai'
|
|
# VOYAGEAI_API_KEY=
|
|
# EMBEDDING_MODEL_PREF='voyage-large-2-instruct'
|
|
|
|
# EMBEDDING_ENGINE='litellm'
|
|
# EMBEDDING_MODEL_PREF='text-embedding-ada-002'
|
|
# EMBEDDING_MODEL_MAX_CHUNK_LENGTH=8192
|
|
# LITE_LLM_BASE_PATH='http://127.0.0.1:4000'
|
|
# LITE_LLM_API_KEY='sk-123abc'
|
|
|
|
###########################################
|
|
######## Vector Database Selection ########
|
|
###########################################
|
|
# Enable all below if you are using vector database: Chroma.
|
|
# VECTOR_DB="chroma"
|
|
# CHROMA_ENDPOINT='http://host.docker.internal:8000'
|
|
# CHROMA_API_HEADER="X-Api-Key"
|
|
# CHROMA_API_KEY="sk-123abc"
|
|
|
|
# Enable all below if you are using vector database: Pinecone.
|
|
# VECTOR_DB="pinecone"
|
|
# PINECONE_API_KEY=
|
|
# PINECONE_INDEX=
|
|
|
|
# Enable all below if you are using vector database: LanceDB.
|
|
# VECTOR_DB="lancedb"
|
|
|
|
# Enable all below if you are using vector database: Weaviate.
|
|
# VECTOR_DB="weaviate"
|
|
# WEAVIATE_ENDPOINT="http://localhost:8080"
|
|
# WEAVIATE_API_KEY=
|
|
|
|
# Enable all below if you are using vector database: Qdrant.
|
|
# VECTOR_DB="qdrant"
|
|
# QDRANT_ENDPOINT="http://localhost:6333"
|
|
# QDRANT_API_KEY=
|
|
|
|
# Enable all below if you are using vector database: Milvus.
|
|
# VECTOR_DB="milvus"
|
|
# MILVUS_ADDRESS="http://localhost:19530"
|
|
# MILVUS_USERNAME=
|
|
# MILVUS_PASSWORD=
|
|
|
|
# Enable all below if you are using vector database: Zilliz Cloud.
|
|
# VECTOR_DB="zilliz"
|
|
# ZILLIZ_ENDPOINT="https://sample.api.gcp-us-west1.zillizcloud.com"
|
|
# ZILLIZ_API_TOKEN=api-token-here
|
|
|
|
# Enable all below if you are using vector database: Astra DB.
|
|
# VECTOR_DB="astra"
|
|
# ASTRA_DB_APPLICATION_TOKEN=
|
|
# ASTRA_DB_ENDPOINT=
|
|
|
|
###########################################
|
|
######## Audio Model Selection ############
|
|
###########################################
|
|
# (default) use built-in whisper-small model.
|
|
# WHISPER_PROVIDER="local"
|
|
|
|
# use openai hosted whisper model.
|
|
# WHISPER_PROVIDER="openai"
|
|
# OPEN_AI_KEY=sk-xxxxxxxx
|
|
|
|
###########################################
|
|
######## TTS/STT Model Selection ##########
|
|
###########################################
|
|
# TTS_PROVIDER="native"
|
|
|
|
# TTS_PROVIDER="openai"
|
|
# TTS_OPEN_AI_KEY=sk-example
|
|
# TTS_OPEN_AI_VOICE_MODEL=nova
|
|
|
|
# TTS_PROVIDER="elevenlabs"
|
|
# TTS_ELEVEN_LABS_KEY=
|
|
# TTS_ELEVEN_LABS_VOICE_MODEL=21m00Tcm4TlvDq8ikWAM # Rachel
|
|
|
|
# CLOUD DEPLOYMENT VARIRABLES ONLY
|
|
# AUTH_TOKEN="hunter2" # This is the password to your application if remote hosting.
|
|
# DISABLE_TELEMETRY="false"
|
|
|
|
###########################################
|
|
######## PASSWORD COMPLEXITY ##############
|
|
###########################################
|
|
# Enforce a password schema for your organization users.
|
|
# Documentation on how to use https://github.com/kamronbatman/joi-password-complexity
|
|
# Default is only 8 char minimum
|
|
# PASSWORDMINCHAR=8
|
|
# PASSWORDMAXCHAR=250
|
|
# PASSWORDLOWERCASE=1
|
|
# PASSWORDUPPERCASE=1
|
|
# PASSWORDNUMERIC=1
|
|
# PASSWORDSYMBOL=1
|
|
# PASSWORDREQUIREMENTS=4
|
|
|
|
###########################################
|
|
######## ENABLE HTTPS SERVER ##############
|
|
###########################################
|
|
# By enabling this and providing the path/filename for the key and cert,
|
|
# the server will use HTTPS instead of HTTP.
|
|
#ENABLE_HTTPS="true"
|
|
#HTTPS_CERT_PATH="sslcert/cert.pem"
|
|
#HTTPS_KEY_PATH="sslcert/key.pem"
|
|
|
|
###########################################
|
|
######## AGENT SERVICE KEYS ###############
|
|
###########################################
|
|
|
|
#------ SEARCH ENGINES -------
|
|
#=============================
|
|
#------ Google Search -------- https://programmablesearchengine.google.com/controlpanel/create
|
|
# AGENT_GSE_KEY=
|
|
# AGENT_GSE_CTX=
|
|
|
|
#------ Serper.dev ----------- https://serper.dev/
|
|
# AGENT_SERPER_DEV_KEY=
|
|
|
|
#------ Bing Search ----------- https://portal.azure.com/
|
|
# AGENT_BING_SEARCH_API_KEY=
|
|
|
|
#------ Serply.io ----------- https://serply.io/
|
|
# AGENT_SERPLY_API_KEY=
|
|
|
|
#------ SearXNG ----------- https://github.com/searxng/searxng
|
|
# AGENT_SEARXNG_API_URL= |