mirror of
https://github.com/Mintplex-Labs/anything-llm.git
synced 2024-11-05 22:40:12 +01:00
88cdd8c872
* Implement use of native embedder (all-Mini-L6-v2) stop showing prisma queries during dev * Add native embedder as an available embedder selection * wrap model loader in try/catch * print progress on download * Update to progress output for embedder * move embedder selection options to component * forgot import * add Data privacy alert updates for local embedder |
||
---|---|---|
.. | ||
.gitignore | ||
README.md |
Native models used by AnythingLLM
This folder is specifically created as a local cache and storage folder that is used for native models that can run on a CPU.
Currently, AnythingLLM uses this folder for the following parts of the application.
Embedding
When your embedding engine preference is native
we will use the ONNX all-MiniLM-L6-v2 model built by Xenova on HuggingFace.co. This model is a quantized and WASM version of the popular all-MiniLM-L6-v2 which produces a 384-dimension vector.
If you are using the native
embedding engine your vector database should be configured to accept 384-dimension models if that parameter is directly editable (Pinecone only).
Text generation (LLM selection)
in progress