anything-llm/server/utils/AiProviders/perplexity/models.js
Sean Hatfield 80ced5eba4
[FEAT] PerplexityAI Support (#778)
* add LLM support for perplexity

* update README & example env

* fix ENV keys in example env files

* slight changes for QA of perplexity support

* Update Perplexity AI name

---------

Co-authored-by: timothycarambat <rambat1010@gmail.com>
2024-02-22 12:48:57 -08:00

50 lines
1.0 KiB
JavaScript

const MODELS = {
"codellama-34b-instruct": {
id: "codellama-34b-instruct",
name: "codellama-34b-instruct",
maxLength: 16384,
},
"codellama-70b-instruct": {
id: "codellama-70b-instruct",
name: "codellama-70b-instruct",
maxLength: 16384,
},
"llama-2-70b-chat": {
id: "llama-2-70b-chat",
name: "llama-2-70b-chat",
maxLength: 4096,
},
"mistral-7b-instruct": {
id: "mistral-7b-instruct",
name: "mistral-7b-instruct",
maxLength: 8192,
},
"mixtral-8x7b-instruct": {
id: "mixtral-8x7b-instruct",
name: "mixtral-8x7b-instruct",
maxLength: 8192,
},
"pplx-7b-chat": {
id: "pplx-7b-chat",
name: "pplx-7b-chat",
maxLength: 8192,
},
"pplx-70b-chat": {
id: "pplx-70b-chat",
name: "pplx-70b-chat",
maxLength: 8192,
},
"pplx-7b-online": {
id: "pplx-7b-online",
name: "pplx-7b-online",
maxLength: 8192,
},
"pplx-70b-online": {
id: "pplx-70b-online",
name: "pplx-70b-online",
maxLength: 8192,
},
};
module.exports.MODELS = MODELS;