mirror of
https://github.com/Mintplex-Labs/anything-llm.git
synced 2025-04-17 18:18:11 +00:00
Bump all static model providers (#1101)
This commit is contained in:
parent
6fde5570b3
commit
8306098b08
8 changed files with 573 additions and 339 deletions
server/utils/AiProviders
openRouter
perplexity
togetherAi
|
@ -1,10 +1,4 @@
|
|||
const MODELS = {
|
||||
"nousresearch/nous-capybara-34b": {
|
||||
id: "nousresearch/nous-capybara-34b",
|
||||
name: "Nous: Capybara 34B",
|
||||
organization: "Nousresearch",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"openrouter/auto": {
|
||||
id: "openrouter/auto",
|
||||
name: "Auto (best for prompt)",
|
||||
|
@ -21,6 +15,12 @@ const MODELS = {
|
|||
id: "mistralai/mistral-7b-instruct:free",
|
||||
name: "Mistral 7B Instruct (free)",
|
||||
organization: "Mistralai",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"openchat/openchat-7b:free": {
|
||||
id: "openchat/openchat-7b:free",
|
||||
name: "OpenChat 3.5 (free)",
|
||||
organization: "Openchat",
|
||||
maxLength: 8192,
|
||||
},
|
||||
"gryphe/mythomist-7b:free": {
|
||||
|
@ -45,13 +45,7 @@ const MODELS = {
|
|||
id: "google/gemma-7b-it:free",
|
||||
name: "Google: Gemma 7B (free)",
|
||||
organization: "Google",
|
||||
maxLength: 8000,
|
||||
},
|
||||
"jondurbin/bagel-34b": {
|
||||
id: "jondurbin/bagel-34b",
|
||||
name: "Bagel 34B v0.2",
|
||||
organization: "Jondurbin",
|
||||
maxLength: 8000,
|
||||
maxLength: 8192,
|
||||
},
|
||||
"jebcarter/psyfighter-13b": {
|
||||
id: "jebcarter/psyfighter-13b",
|
||||
|
@ -65,54 +59,12 @@ const MODELS = {
|
|||
organization: "Koboldai",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"neversleep/noromaid-mixtral-8x7b-instruct": {
|
||||
id: "neversleep/noromaid-mixtral-8x7b-instruct",
|
||||
name: "Noromaid Mixtral 8x7B Instruct",
|
||||
organization: "Neversleep",
|
||||
maxLength: 8000,
|
||||
},
|
||||
"nousresearch/nous-hermes-llama2-13b": {
|
||||
id: "nousresearch/nous-hermes-llama2-13b",
|
||||
name: "Nous: Hermes 13B",
|
||||
organization: "Nousresearch",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"meta-llama/codellama-34b-instruct": {
|
||||
id: "meta-llama/codellama-34b-instruct",
|
||||
name: "Meta: CodeLlama 34B Instruct",
|
||||
organization: "Meta-llama",
|
||||
maxLength: 8192,
|
||||
},
|
||||
"phind/phind-codellama-34b": {
|
||||
id: "phind/phind-codellama-34b",
|
||||
name: "Phind: CodeLlama 34B v2",
|
||||
organization: "Phind",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"intel/neural-chat-7b": {
|
||||
id: "intel/neural-chat-7b",
|
||||
name: "Neural Chat 7B v3.1",
|
||||
organization: "Intel",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"mistralai/mixtral-8x7b-instruct": {
|
||||
id: "mistralai/mixtral-8x7b-instruct",
|
||||
name: "Mistral: Mixtral 8x7B Instruct",
|
||||
organization: "Mistralai",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"nousresearch/nous-hermes-2-mixtral-8x7b-dpo": {
|
||||
id: "nousresearch/nous-hermes-2-mixtral-8x7b-dpo",
|
||||
name: "Nous: Hermes 2 Mixtral 8x7B DPO",
|
||||
organization: "Nousresearch",
|
||||
maxLength: 32000,
|
||||
},
|
||||
"nousresearch/nous-hermes-2-mixtral-8x7b-sft": {
|
||||
id: "nousresearch/nous-hermes-2-mixtral-8x7b-sft",
|
||||
name: "Nous: Hermes 2 Mixtral 8x7B SFT",
|
||||
organization: "Nousresearch",
|
||||
maxLength: 32000,
|
||||
},
|
||||
"haotian-liu/llava-13b": {
|
||||
id: "haotian-liu/llava-13b",
|
||||
name: "Llava 13B",
|
||||
|
@ -143,30 +95,12 @@ const MODELS = {
|
|||
organization: "Pygmalionai",
|
||||
maxLength: 8192,
|
||||
},
|
||||
"undi95/remm-slerp-l2-13b-6k": {
|
||||
id: "undi95/remm-slerp-l2-13b-6k",
|
||||
name: "ReMM SLERP 13B 6k",
|
||||
organization: "Undi95",
|
||||
maxLength: 6144,
|
||||
},
|
||||
"gryphe/mythomax-l2-13b": {
|
||||
id: "gryphe/mythomax-l2-13b",
|
||||
name: "MythoMax 13B",
|
||||
organization: "Gryphe",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"xwin-lm/xwin-lm-70b": {
|
||||
id: "xwin-lm/xwin-lm-70b",
|
||||
name: "Xwin 70B",
|
||||
organization: "Xwin-lm",
|
||||
maxLength: 8192,
|
||||
},
|
||||
"gryphe/mythomax-l2-13b-8k": {
|
||||
id: "gryphe/mythomax-l2-13b-8k",
|
||||
name: "MythoMax 13B 8k",
|
||||
organization: "Gryphe",
|
||||
maxLength: 8192,
|
||||
},
|
||||
"alpindale/goliath-120b": {
|
||||
id: "alpindale/goliath-120b",
|
||||
name: "Goliath 120B",
|
||||
|
@ -185,15 +119,27 @@ const MODELS = {
|
|||
organization: "Gryphe",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"sophosympatheia/midnight-rose-70b": {
|
||||
id: "sophosympatheia/midnight-rose-70b",
|
||||
name: "Midnight Rose 70B",
|
||||
organization: "Sophosympatheia",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"undi95/remm-slerp-l2-13b:extended": {
|
||||
id: "undi95/remm-slerp-l2-13b:extended",
|
||||
name: "ReMM SLERP 13B (extended)",
|
||||
organization: "Undi95",
|
||||
maxLength: 6144,
|
||||
},
|
||||
"mancer/weaver": {
|
||||
id: "mancer/weaver",
|
||||
name: "Mancer: Weaver (alpha)",
|
||||
organization: "Mancer",
|
||||
maxLength: 8000,
|
||||
},
|
||||
"nousresearch/nous-hermes-llama2-70b": {
|
||||
id: "nousresearch/nous-hermes-llama2-70b",
|
||||
name: "Nous: Hermes 70B",
|
||||
"nousresearch/nous-hermes-llama2-13b": {
|
||||
id: "nousresearch/nous-hermes-llama2-13b",
|
||||
name: "Nous: Hermes 13B",
|
||||
organization: "Nousresearch",
|
||||
maxLength: 4096,
|
||||
},
|
||||
|
@ -203,12 +149,24 @@ const MODELS = {
|
|||
organization: "Nousresearch",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"meta-llama/codellama-34b-instruct": {
|
||||
id: "meta-llama/codellama-34b-instruct",
|
||||
name: "Meta: CodeLlama 34B Instruct",
|
||||
organization: "Meta-llama",
|
||||
maxLength: 8192,
|
||||
},
|
||||
"codellama/codellama-70b-instruct": {
|
||||
id: "codellama/codellama-70b-instruct",
|
||||
name: "Meta: CodeLlama 70B Instruct",
|
||||
organization: "Codellama",
|
||||
maxLength: 2048,
|
||||
},
|
||||
"phind/phind-codellama-34b": {
|
||||
id: "phind/phind-codellama-34b",
|
||||
name: "Phind: CodeLlama 34B v2",
|
||||
organization: "Phind",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"teknium/openhermes-2-mistral-7b": {
|
||||
id: "teknium/openhermes-2-mistral-7b",
|
||||
name: "OpenHermes 2 Mistral 7B",
|
||||
|
@ -227,12 +185,6 @@ const MODELS = {
|
|||
organization: "Undi95",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"undi95/toppy-m-7b": {
|
||||
id: "undi95/toppy-m-7b",
|
||||
name: "Toppy M 7B",
|
||||
organization: "Undi95",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"openrouter/cinematika-7b": {
|
||||
id: "openrouter/cinematika-7b",
|
||||
name: "Cinematika 7B (alpha)",
|
||||
|
@ -271,7 +223,7 @@ const MODELS = {
|
|||
},
|
||||
"mistralai/mixtral-8x7b": {
|
||||
id: "mistralai/mixtral-8x7b",
|
||||
name: "Mistral: Mixtral 8x7B (base)",
|
||||
name: "Mixtral 8x7B (base)",
|
||||
organization: "Mistralai",
|
||||
maxLength: 32768,
|
||||
},
|
||||
|
@ -281,6 +233,12 @@ const MODELS = {
|
|||
organization: "Nousresearch",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"nousresearch/nous-hermes-2-mixtral-8x7b-sft": {
|
||||
id: "nousresearch/nous-hermes-2-mixtral-8x7b-sft",
|
||||
name: "Nous: Hermes 2 Mixtral 8x7B SFT",
|
||||
organization: "Nousresearch",
|
||||
maxLength: 32000,
|
||||
},
|
||||
"nousresearch/nous-hermes-2-mistral-7b-dpo": {
|
||||
id: "nousresearch/nous-hermes-2-mistral-7b-dpo",
|
||||
name: "Nous: Hermes 2 Mistral 7B DPO",
|
||||
|
@ -303,7 +261,7 @@ const MODELS = {
|
|||
id: "openai/gpt-3.5-turbo",
|
||||
name: "OpenAI: GPT-3.5 Turbo",
|
||||
organization: "Openai",
|
||||
maxLength: 4095,
|
||||
maxLength: 16385,
|
||||
},
|
||||
"openai/gpt-3.5-turbo-0125": {
|
||||
id: "openai/gpt-3.5-turbo-0125",
|
||||
|
@ -335,9 +293,15 @@ const MODELS = {
|
|||
organization: "Openai",
|
||||
maxLength: 16385,
|
||||
},
|
||||
"openai/gpt-4-turbo": {
|
||||
id: "openai/gpt-4-turbo",
|
||||
name: "OpenAI: GPT-4 Turbo",
|
||||
organization: "Openai",
|
||||
maxLength: 128000,
|
||||
},
|
||||
"openai/gpt-4-turbo-preview": {
|
||||
id: "openai/gpt-4-turbo-preview",
|
||||
name: "OpenAI: GPT-4 Turbo (preview)",
|
||||
name: "OpenAI: GPT-4 Turbo Preview",
|
||||
organization: "Openai",
|
||||
maxLength: 128000,
|
||||
},
|
||||
|
@ -373,7 +337,7 @@ const MODELS = {
|
|||
},
|
||||
"openai/gpt-4-vision-preview": {
|
||||
id: "openai/gpt-4-vision-preview",
|
||||
name: "OpenAI: GPT-4 Vision (preview)",
|
||||
name: "OpenAI: GPT-4 Vision",
|
||||
organization: "Openai",
|
||||
maxLength: 128000,
|
||||
},
|
||||
|
@ -387,37 +351,43 @@ const MODELS = {
|
|||
id: "google/palm-2-chat-bison",
|
||||
name: "Google: PaLM 2 Chat",
|
||||
organization: "Google",
|
||||
maxLength: 36864,
|
||||
maxLength: 25804,
|
||||
},
|
||||
"google/palm-2-codechat-bison": {
|
||||
id: "google/palm-2-codechat-bison",
|
||||
name: "Google: PaLM 2 Code Chat",
|
||||
organization: "Google",
|
||||
maxLength: 28672,
|
||||
maxLength: 20070,
|
||||
},
|
||||
"google/palm-2-chat-bison-32k": {
|
||||
id: "google/palm-2-chat-bison-32k",
|
||||
name: "Google: PaLM 2 Chat 32k",
|
||||
organization: "Google",
|
||||
maxLength: 131072,
|
||||
maxLength: 91750,
|
||||
},
|
||||
"google/palm-2-codechat-bison-32k": {
|
||||
id: "google/palm-2-codechat-bison-32k",
|
||||
name: "Google: PaLM 2 Code Chat 32k",
|
||||
organization: "Google",
|
||||
maxLength: 131072,
|
||||
maxLength: 91750,
|
||||
},
|
||||
"google/gemini-pro": {
|
||||
id: "google/gemini-pro",
|
||||
name: "Google: Gemini Pro (preview)",
|
||||
name: "Google: Gemini Pro 1.0",
|
||||
organization: "Google",
|
||||
maxLength: 131040,
|
||||
maxLength: 91728,
|
||||
},
|
||||
"google/gemini-pro-vision": {
|
||||
id: "google/gemini-pro-vision",
|
||||
name: "Google: Gemini Pro Vision (preview)",
|
||||
name: "Google: Gemini Pro Vision 1.0",
|
||||
organization: "Google",
|
||||
maxLength: 65536,
|
||||
maxLength: 45875,
|
||||
},
|
||||
"google/gemini-pro-1.5": {
|
||||
id: "google/gemini-pro-1.5",
|
||||
name: "Google: Gemini Pro 1.5 (preview)",
|
||||
organization: "Google",
|
||||
maxLength: 2800000,
|
||||
},
|
||||
"perplexity/pplx-70b-online": {
|
||||
id: "perplexity/pplx-70b-online",
|
||||
|
@ -443,18 +413,96 @@ const MODELS = {
|
|||
organization: "Perplexity",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"perplexity/sonar-small-chat": {
|
||||
id: "perplexity/sonar-small-chat",
|
||||
name: "Perplexity: Sonar 7B",
|
||||
organization: "Perplexity",
|
||||
maxLength: 16384,
|
||||
},
|
||||
"perplexity/sonar-medium-chat": {
|
||||
id: "perplexity/sonar-medium-chat",
|
||||
name: "Perplexity: Sonar 8x7B",
|
||||
organization: "Perplexity",
|
||||
maxLength: 16384,
|
||||
},
|
||||
"perplexity/sonar-small-online": {
|
||||
id: "perplexity/sonar-small-online",
|
||||
name: "Perplexity: Sonar 7B Online",
|
||||
organization: "Perplexity",
|
||||
maxLength: 12000,
|
||||
},
|
||||
"perplexity/sonar-medium-online": {
|
||||
id: "perplexity/sonar-medium-online",
|
||||
name: "Perplexity: Sonar 8x7B Online",
|
||||
organization: "Perplexity",
|
||||
maxLength: 12000,
|
||||
},
|
||||
"fireworks/mixtral-8x22b-instruct-preview": {
|
||||
id: "fireworks/mixtral-8x22b-instruct-preview",
|
||||
name: "Fireworks Mixtral 8x22B Instruct OH (preview)",
|
||||
organization: "Fireworks",
|
||||
maxLength: 8192,
|
||||
},
|
||||
"anthropic/claude-3-opus": {
|
||||
id: "anthropic/claude-3-opus",
|
||||
name: "Anthropic: Claude 3 Opus",
|
||||
organization: "Anthropic",
|
||||
maxLength: 200000,
|
||||
},
|
||||
"anthropic/claude-3-sonnet": {
|
||||
id: "anthropic/claude-3-sonnet",
|
||||
name: "Anthropic: Claude 3 Sonnet",
|
||||
organization: "Anthropic",
|
||||
maxLength: 200000,
|
||||
},
|
||||
"anthropic/claude-3-haiku": {
|
||||
id: "anthropic/claude-3-haiku",
|
||||
name: "Anthropic: Claude 3 Haiku",
|
||||
organization: "Anthropic",
|
||||
maxLength: 200000,
|
||||
},
|
||||
"anthropic/claude-3-opus:beta": {
|
||||
id: "anthropic/claude-3-opus:beta",
|
||||
name: "Anthropic: Claude 3 Opus (self-moderated)",
|
||||
organization: "Anthropic",
|
||||
maxLength: 200000,
|
||||
},
|
||||
"anthropic/claude-3-sonnet:beta": {
|
||||
id: "anthropic/claude-3-sonnet:beta",
|
||||
name: "Anthropic: Claude 3 Sonnet (self-moderated)",
|
||||
organization: "Anthropic",
|
||||
maxLength: 200000,
|
||||
},
|
||||
"anthropic/claude-3-haiku:beta": {
|
||||
id: "anthropic/claude-3-haiku:beta",
|
||||
name: "Anthropic: Claude 3 Haiku (self-moderated)",
|
||||
organization: "Anthropic",
|
||||
maxLength: 200000,
|
||||
},
|
||||
"meta-llama/llama-2-70b-chat": {
|
||||
id: "meta-llama/llama-2-70b-chat",
|
||||
name: "Meta: Llama v2 70B Chat",
|
||||
organization: "Meta-llama",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"nousresearch/nous-capybara-34b": {
|
||||
id: "nousresearch/nous-capybara-34b",
|
||||
name: "Nous: Capybara 34B",
|
||||
organization: "Nousresearch",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"jondurbin/airoboros-l2-70b": {
|
||||
id: "jondurbin/airoboros-l2-70b",
|
||||
name: "Airoboros 70B",
|
||||
organization: "Jondurbin",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"jondurbin/bagel-34b": {
|
||||
id: "jondurbin/bagel-34b",
|
||||
name: "Bagel 34B v0.2",
|
||||
organization: "Jondurbin",
|
||||
maxLength: 8000,
|
||||
},
|
||||
"austism/chronos-hermes-13b": {
|
||||
id: "austism/chronos-hermes-13b",
|
||||
name: "Chronos Hermes 13B v2",
|
||||
|
@ -465,7 +513,13 @@ const MODELS = {
|
|||
id: "mistralai/mistral-7b-instruct",
|
||||
name: "Mistral 7B Instruct",
|
||||
organization: "Mistralai",
|
||||
maxLength: 8192,
|
||||
maxLength: 32768,
|
||||
},
|
||||
"gryphe/mythomax-l2-13b": {
|
||||
id: "gryphe/mythomax-l2-13b",
|
||||
name: "MythoMax 13B",
|
||||
organization: "Gryphe",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"openchat/openchat-7b": {
|
||||
id: "openchat/openchat-7b",
|
||||
|
@ -473,18 +527,42 @@ const MODELS = {
|
|||
organization: "Openchat",
|
||||
maxLength: 8192,
|
||||
},
|
||||
"undi95/toppy-m-7b": {
|
||||
id: "undi95/toppy-m-7b",
|
||||
name: "Toppy M 7B",
|
||||
organization: "Undi95",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"lizpreciatior/lzlv-70b-fp16-hf": {
|
||||
id: "lizpreciatior/lzlv-70b-fp16-hf",
|
||||
name: "lzlv 70B",
|
||||
organization: "Lizpreciatior",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"mistralai/mixtral-8x7b-instruct": {
|
||||
id: "mistralai/mixtral-8x7b-instruct",
|
||||
name: "Mixtral 8x7B Instruct",
|
||||
organization: "Mistralai",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"cognitivecomputations/dolphin-mixtral-8x7b": {
|
||||
id: "cognitivecomputations/dolphin-mixtral-8x7b",
|
||||
name: "Dolphin 2.6 Mixtral 8x7B 🐬",
|
||||
organization: "Cognitivecomputations",
|
||||
maxLength: 32000,
|
||||
},
|
||||
"neversleep/noromaid-mixtral-8x7b-instruct": {
|
||||
id: "neversleep/noromaid-mixtral-8x7b-instruct",
|
||||
name: "Noromaid Mixtral 8x7B Instruct",
|
||||
organization: "Neversleep",
|
||||
maxLength: 8000,
|
||||
},
|
||||
"nousresearch/nous-hermes-2-mixtral-8x7b-dpo": {
|
||||
id: "nousresearch/nous-hermes-2-mixtral-8x7b-dpo",
|
||||
name: "Nous: Hermes 2 Mixtral 8x7B DPO",
|
||||
organization: "Nousresearch",
|
||||
maxLength: 32000,
|
||||
},
|
||||
"rwkv/rwkv-5-world-3b": {
|
||||
id: "rwkv/rwkv-5-world-3b",
|
||||
name: "RWKV v5 World 3B",
|
||||
|
@ -507,7 +585,19 @@ const MODELS = {
|
|||
id: "google/gemma-7b-it",
|
||||
name: "Google: Gemma 7B",
|
||||
organization: "Google",
|
||||
maxLength: 8000,
|
||||
maxLength: 8192,
|
||||
},
|
||||
"databricks/dbrx-instruct": {
|
||||
id: "databricks/dbrx-instruct",
|
||||
name: "Databricks: DBRX 132B Instruct",
|
||||
organization: "Databricks",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"huggingfaceh4/zephyr-orpo-141b-a35b": {
|
||||
id: "huggingfaceh4/zephyr-orpo-141b-a35b",
|
||||
name: "Zephyr 141B-A35B",
|
||||
organization: "Huggingfaceh4",
|
||||
maxLength: 65536,
|
||||
},
|
||||
"anthropic/claude-2": {
|
||||
id: "anthropic/claude-2",
|
||||
|
@ -565,58 +655,124 @@ const MODELS = {
|
|||
},
|
||||
"anthropic/claude-2:beta": {
|
||||
id: "anthropic/claude-2:beta",
|
||||
name: "Anthropic: Claude v2 (experimental)",
|
||||
name: "Anthropic: Claude v2 (self-moderated)",
|
||||
organization: "Anthropic",
|
||||
maxLength: 200000,
|
||||
},
|
||||
"anthropic/claude-2.1:beta": {
|
||||
id: "anthropic/claude-2.1:beta",
|
||||
name: "Anthropic: Claude v2.1 (experimental)",
|
||||
name: "Anthropic: Claude v2.1 (self-moderated)",
|
||||
organization: "Anthropic",
|
||||
maxLength: 200000,
|
||||
},
|
||||
"anthropic/claude-2.0:beta": {
|
||||
id: "anthropic/claude-2.0:beta",
|
||||
name: "Anthropic: Claude v2.0 (experimental)",
|
||||
name: "Anthropic: Claude v2.0 (self-moderated)",
|
||||
organization: "Anthropic",
|
||||
maxLength: 100000,
|
||||
},
|
||||
"anthropic/claude-instant-1:beta": {
|
||||
id: "anthropic/claude-instant-1:beta",
|
||||
name: "Anthropic: Claude Instant v1 (experimental)",
|
||||
name: "Anthropic: Claude Instant v1 (self-moderated)",
|
||||
organization: "Anthropic",
|
||||
maxLength: 100000,
|
||||
},
|
||||
"mistralai/mixtral-8x22b": {
|
||||
id: "mistralai/mixtral-8x22b",
|
||||
name: "Mistral: Mixtral 8x22B (base)",
|
||||
organization: "Mistralai",
|
||||
maxLength: 65536,
|
||||
},
|
||||
"huggingfaceh4/zephyr-7b-beta:free": {
|
||||
id: "huggingfaceh4/zephyr-7b-beta:free",
|
||||
name: "Hugging Face: Zephyr 7B (free)",
|
||||
organization: "Huggingfaceh4",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"openchat/openchat-7b:free": {
|
||||
id: "openchat/openchat-7b:free",
|
||||
name: "OpenChat 3.5 (free)",
|
||||
organization: "Openchat",
|
||||
"mistralai/mixtral-8x7b-instruct:nitro": {
|
||||
id: "mistralai/mixtral-8x7b-instruct:nitro",
|
||||
name: "Mixtral 8x7B Instruct (nitro)",
|
||||
organization: "Mistralai",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"meta-llama/llama-2-70b-chat:nitro": {
|
||||
id: "meta-llama/llama-2-70b-chat:nitro",
|
||||
name: "Meta: Llama v2 70B Chat (nitro)",
|
||||
organization: "Meta-llama",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"gryphe/mythomax-l2-13b:nitro": {
|
||||
id: "gryphe/mythomax-l2-13b:nitro",
|
||||
name: "MythoMax 13B (nitro)",
|
||||
organization: "Gryphe",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"mistralai/mistral-7b-instruct:nitro": {
|
||||
id: "mistralai/mistral-7b-instruct:nitro",
|
||||
name: "Mistral 7B Instruct (nitro)",
|
||||
organization: "Mistralai",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"google/gemma-7b-it:nitro": {
|
||||
id: "google/gemma-7b-it:nitro",
|
||||
name: "Google: Gemma 7B (nitro)",
|
||||
organization: "Google",
|
||||
maxLength: 8192,
|
||||
},
|
||||
"databricks/dbrx-instruct:nitro": {
|
||||
id: "databricks/dbrx-instruct:nitro",
|
||||
name: "Databricks: DBRX 132B Instruct (nitro)",
|
||||
organization: "Databricks",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"gryphe/mythomax-l2-13b:extended": {
|
||||
id: "gryphe/mythomax-l2-13b:extended",
|
||||
name: "MythoMax 13B (extended)",
|
||||
organization: "Gryphe",
|
||||
maxLength: 8192,
|
||||
},
|
||||
"mistralai/mistral-tiny": {
|
||||
id: "mistralai/mistral-tiny",
|
||||
name: "Mistral: Tiny",
|
||||
name: "Mistral Tiny",
|
||||
organization: "Mistralai",
|
||||
maxLength: 32000,
|
||||
},
|
||||
"mistralai/mistral-small": {
|
||||
id: "mistralai/mistral-small",
|
||||
name: "Mistral: Small",
|
||||
name: "Mistral Small",
|
||||
organization: "Mistralai",
|
||||
maxLength: 32000,
|
||||
},
|
||||
"mistralai/mistral-medium": {
|
||||
id: "mistralai/mistral-medium",
|
||||
name: "Mistral: Medium",
|
||||
name: "Mistral Medium",
|
||||
organization: "Mistralai",
|
||||
maxLength: 32000,
|
||||
},
|
||||
"mistralai/mistral-large": {
|
||||
id: "mistralai/mistral-large",
|
||||
name: "Mistral Large",
|
||||
organization: "Mistralai",
|
||||
maxLength: 32000,
|
||||
},
|
||||
"cohere/command": {
|
||||
id: "cohere/command",
|
||||
name: "Cohere: Command",
|
||||
organization: "Cohere",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"cohere/command-r": {
|
||||
id: "cohere/command-r",
|
||||
name: "Cohere: Command R",
|
||||
organization: "Cohere",
|
||||
maxLength: 128000,
|
||||
},
|
||||
"cohere/command-r-plus": {
|
||||
id: "cohere/command-r-plus",
|
||||
name: "Cohere: Command R+",
|
||||
organization: "Cohere",
|
||||
maxLength: 128000,
|
||||
},
|
||||
};
|
||||
|
||||
module.exports.MODELS = MODELS;
|
||||
|
|
|
@ -6,7 +6,7 @@
|
|||
// copy outputs into the export in ../models.js
|
||||
|
||||
// Update the date below if you run this again because OpenRouter added new models.
|
||||
// Last Collected: Feb 23, 2024
|
||||
// Last Collected: Apr 14, 2024
|
||||
|
||||
import fs from "fs";
|
||||
|
||||
|
|
|
@ -19,21 +19,11 @@ const MODELS = {
|
|||
name: "sonar-medium-online",
|
||||
maxLength: 12000,
|
||||
},
|
||||
"codellama-34b-instruct": {
|
||||
id: "codellama-34b-instruct",
|
||||
name: "codellama-34b-instruct",
|
||||
maxLength: 16384,
|
||||
},
|
||||
"codellama-70b-instruct": {
|
||||
id: "codellama-70b-instruct",
|
||||
name: "codellama-70b-instruct",
|
||||
maxLength: 16384,
|
||||
},
|
||||
"llama-2-70b-chat": {
|
||||
id: "llama-2-70b-chat",
|
||||
name: "llama-2-70b-chat",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"mistral-7b-instruct": {
|
||||
id: "mistral-7b-instruct",
|
||||
name: "mistral-7b-instruct",
|
||||
|
@ -44,26 +34,6 @@ const MODELS = {
|
|||
name: "mixtral-8x7b-instruct",
|
||||
maxLength: 16384,
|
||||
},
|
||||
"pplx-7b-chat": {
|
||||
id: "pplx-7b-chat",
|
||||
name: "pplx-7b-chat",
|
||||
maxLength: 16384,
|
||||
},
|
||||
"pplx-7b-online": {
|
||||
id: "pplx-7b-online",
|
||||
name: "pplx-7b-online",
|
||||
maxLength: 12000,
|
||||
},
|
||||
"pplx-70b-chat": {
|
||||
id: "pplx-70b-chat",
|
||||
name: "pplx-70b-chat",
|
||||
maxLength: 8192,
|
||||
},
|
||||
"pplx-70b-online": {
|
||||
id: "pplx-70b-online",
|
||||
name: "pplx-70b-online",
|
||||
maxLength: 4000,
|
||||
},
|
||||
};
|
||||
|
||||
module.exports.MODELS = MODELS;
|
||||
|
|
|
@ -1,15 +1,9 @@
|
|||
| Model | Parameter Count | Context Length | Model Type |
|
||||
| :-------------------------- | :-------------- | :------------- | :-------------- |
|
||||
| `sonar-small-chat` | 7B | 16384 | Chat Completion |
|
||||
| `sonar-small-online` | 7B | 12000 | Chat Completion |
|
||||
| `sonar-medium-chat` | 8x7B | 16384 | Chat Completion |
|
||||
| `sonar-medium-online` | 8x7B | 12000 | Chat Completion |
|
||||
| `codellama-34b-instruct`[3] | 34B | 16384 | Chat Completion |
|
||||
| `codellama-70b-instruct` | 70B | 16384 | Chat Completion |
|
||||
| `llama-2-70b-chat`[3] | 70B | 4096 | Chat Completion |
|
||||
| `mistral-7b-instruct` [1] | 7B | 16384 | Chat Completion |
|
||||
| `mixtral-8x7b-instruct` | 8x7B | 16384 | Chat Completion |
|
||||
| `pplx-7b-chat`[2] [3] | 7B | 16384 | Chat Completion |
|
||||
| `pplx-7b-online`[2] [3] | 7B | 12000 | Chat Completion |
|
||||
| `pplx-70b-chat`[3] | 70B | 8192 | Chat Completion |
|
||||
| `pplx-70b-online`[3] | 70B | 4000 | Chat Completion |
|
||||
| Model | Parameter Count | Context Length | Model Type |
|
||||
| :-------------------- | :-------------- | :------------- | :-------------- |
|
||||
| `sonar-small-chat` | 7B | 16384 | Chat Completion |
|
||||
| `sonar-small-online` | 7B | 12000 | Chat Completion |
|
||||
| `sonar-medium-chat` | 8x7B | 16384 | Chat Completion |
|
||||
| `sonar-medium-online` | 8x7B | 12000 | Chat Completion |
|
||||
| `codellama-70b-instruct` | 70B | 16384 | Chat Completion |
|
||||
| `mistral-7b-instruct` [1] | 7B | 16384 | Chat Completion |
|
||||
| `mixtral-8x7b-instruct` | 8x7B | 16384 | Chat Completion |
|
|
@ -8,7 +8,7 @@
|
|||
// copy outputs into the export in ../models.js
|
||||
|
||||
// Update the date below if you run this again because Perplexity added new models.
|
||||
// Last Collected: Feb 23, 2024
|
||||
// Last Collected: Apr 14, 2024
|
||||
|
||||
import fs from "fs";
|
||||
|
||||
|
|
|
@ -1,8 +1,26 @@
|
|||
const MODELS = {
|
||||
"togethercomputer/alpaca-7b": {
|
||||
id: "togethercomputer/alpaca-7b",
|
||||
organization: "Stanford",
|
||||
name: "Alpaca (7B)",
|
||||
"zero-one-ai/Yi-34B-Chat": {
|
||||
id: "zero-one-ai/Yi-34B-Chat",
|
||||
organization: "01.AI",
|
||||
name: "01-ai Yi Chat (34B)",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"allenai/OLMo-7B-Instruct": {
|
||||
id: "allenai/OLMo-7B-Instruct",
|
||||
organization: "Allen AI",
|
||||
name: "OLMo Instruct (7B)",
|
||||
maxLength: 2048,
|
||||
},
|
||||
"allenai/OLMo-7B-Twin-2T": {
|
||||
id: "allenai/OLMo-7B-Twin-2T",
|
||||
organization: "Allen AI",
|
||||
name: "OLMo Twin-2T (7B)",
|
||||
maxLength: 2048,
|
||||
},
|
||||
"allenai/OLMo-7B": {
|
||||
id: "allenai/OLMo-7B",
|
||||
organization: "Allen AI",
|
||||
name: "OLMo (7B)",
|
||||
maxLength: 2048,
|
||||
},
|
||||
"Austism/chronos-hermes-13b": {
|
||||
|
@ -11,96 +29,150 @@ const MODELS = {
|
|||
name: "Chronos Hermes (13B)",
|
||||
maxLength: 2048,
|
||||
},
|
||||
"togethercomputer/CodeLlama-13b-Instruct": {
|
||||
id: "togethercomputer/CodeLlama-13b-Instruct",
|
||||
organization: "Meta",
|
||||
name: "Code Llama Instruct (13B)",
|
||||
"cognitivecomputations/dolphin-2.5-mixtral-8x7b": {
|
||||
id: "cognitivecomputations/dolphin-2.5-mixtral-8x7b",
|
||||
organization: "cognitivecomputations",
|
||||
name: "Dolphin 2.5 Mixtral 8x7b",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"databricks/dbrx-instruct": {
|
||||
id: "databricks/dbrx-instruct",
|
||||
organization: "databricks",
|
||||
name: "DBRX Instruct",
|
||||
maxLength: 32000,
|
||||
},
|
||||
"deepseek-ai/deepseek-coder-33b-instruct": {
|
||||
id: "deepseek-ai/deepseek-coder-33b-instruct",
|
||||
organization: "DeepSeek",
|
||||
name: "Deepseek Coder Instruct (33B)",
|
||||
maxLength: 16384,
|
||||
},
|
||||
"deepseek-ai/deepseek-llm-67b-chat": {
|
||||
id: "deepseek-ai/deepseek-llm-67b-chat",
|
||||
organization: "DeepSeek",
|
||||
name: "DeepSeek LLM Chat (67B)",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"garage-bAInd/Platypus2-70B-instruct": {
|
||||
id: "garage-bAInd/Platypus2-70B-instruct",
|
||||
organization: "garage-bAInd",
|
||||
name: "Platypus2 Instruct (70B)",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"google/gemma-2b-it": {
|
||||
id: "google/gemma-2b-it",
|
||||
organization: "Google",
|
||||
name: "Gemma Instruct (2B)",
|
||||
maxLength: 8192,
|
||||
},
|
||||
"togethercomputer/CodeLlama-34b-Instruct": {
|
||||
id: "togethercomputer/CodeLlama-34b-Instruct",
|
||||
organization: "Meta",
|
||||
name: "Code Llama Instruct (34B)",
|
||||
"google/gemma-7b-it": {
|
||||
id: "google/gemma-7b-it",
|
||||
organization: "Google",
|
||||
name: "Gemma Instruct (7B)",
|
||||
maxLength: 8192,
|
||||
},
|
||||
"togethercomputer/CodeLlama-7b-Instruct": {
|
||||
id: "togethercomputer/CodeLlama-7b-Instruct",
|
||||
organization: "Meta",
|
||||
name: "Code Llama Instruct (7B)",
|
||||
maxLength: 8192,
|
||||
},
|
||||
"DiscoResearch/DiscoLM-mixtral-8x7b-v2": {
|
||||
id: "DiscoResearch/DiscoLM-mixtral-8x7b-v2",
|
||||
organization: "DiscoResearch",
|
||||
name: "DiscoLM Mixtral 8x7b",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"togethercomputer/falcon-40b-instruct": {
|
||||
id: "togethercomputer/falcon-40b-instruct",
|
||||
organization: "TII UAE",
|
||||
name: "Falcon Instruct (40B)",
|
||||
maxLength: 2048,
|
||||
},
|
||||
"togethercomputer/falcon-7b-instruct": {
|
||||
id: "togethercomputer/falcon-7b-instruct",
|
||||
organization: "TII UAE",
|
||||
name: "Falcon Instruct (7B)",
|
||||
maxLength: 2048,
|
||||
},
|
||||
"togethercomputer/GPT-NeoXT-Chat-Base-20B": {
|
||||
id: "togethercomputer/GPT-NeoXT-Chat-Base-20B",
|
||||
organization: "Together",
|
||||
name: "GPT-NeoXT-Chat-Base (20B)",
|
||||
maxLength: 2048,
|
||||
},
|
||||
"togethercomputer/llama-2-13b-chat": {
|
||||
id: "togethercomputer/llama-2-13b-chat",
|
||||
organization: "Meta",
|
||||
name: "LLaMA-2 Chat (13B)",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"togethercomputer/llama-2-70b-chat": {
|
||||
id: "togethercomputer/llama-2-70b-chat",
|
||||
organization: "Meta",
|
||||
name: "LLaMA-2 Chat (70B)",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"togethercomputer/llama-2-7b-chat": {
|
||||
id: "togethercomputer/llama-2-7b-chat",
|
||||
organization: "Meta",
|
||||
name: "LLaMA-2 Chat (7B)",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"togethercomputer/Llama-2-7B-32K-Instruct": {
|
||||
id: "togethercomputer/Llama-2-7B-32K-Instruct",
|
||||
organization: "Together",
|
||||
name: "LLaMA-2-7B-32K-Instruct (7B)",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"mistralai/Mistral-7B-Instruct-v0.1": {
|
||||
id: "mistralai/Mistral-7B-Instruct-v0.1",
|
||||
organization: "MistralAI",
|
||||
name: "Mistral (7B) Instruct v0.1",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"mistralai/Mistral-7B-Instruct-v0.2": {
|
||||
id: "mistralai/Mistral-7B-Instruct-v0.2",
|
||||
organization: "MistralAI",
|
||||
name: "Mistral (7B) Instruct v0.2",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"mistralai/Mixtral-8x7B-Instruct-v0.1": {
|
||||
id: "mistralai/Mixtral-8x7B-Instruct-v0.1",
|
||||
organization: "MistralAI",
|
||||
name: "Mixtral-8x7B Instruct",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"Gryphe/MythoMax-L2-13b": {
|
||||
id: "Gryphe/MythoMax-L2-13b",
|
||||
organization: "Gryphe",
|
||||
name: "MythoMax-L2 (13B)",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"lmsys/vicuna-13b-v1.5": {
|
||||
id: "lmsys/vicuna-13b-v1.5",
|
||||
organization: "LM Sys",
|
||||
name: "Vicuna v1.5 (13B)",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"lmsys/vicuna-7b-v1.5": {
|
||||
id: "lmsys/vicuna-7b-v1.5",
|
||||
organization: "LM Sys",
|
||||
name: "Vicuna v1.5 (7B)",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"codellama/CodeLlama-13b-Instruct-hf": {
|
||||
id: "codellama/CodeLlama-13b-Instruct-hf",
|
||||
organization: "Meta",
|
||||
name: "Code Llama Instruct (13B)",
|
||||
maxLength: 16384,
|
||||
},
|
||||
"codellama/CodeLlama-34b-Instruct-hf": {
|
||||
id: "codellama/CodeLlama-34b-Instruct-hf",
|
||||
organization: "Meta",
|
||||
name: "Code Llama Instruct (34B)",
|
||||
maxLength: 16384,
|
||||
},
|
||||
"codellama/CodeLlama-70b-Instruct-hf": {
|
||||
id: "codellama/CodeLlama-70b-Instruct-hf",
|
||||
organization: "Meta",
|
||||
name: "Code Llama Instruct (70B)",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"codellama/CodeLlama-7b-Instruct-hf": {
|
||||
id: "codellama/CodeLlama-7b-Instruct-hf",
|
||||
organization: "Meta",
|
||||
name: "Code Llama Instruct (7B)",
|
||||
maxLength: 16384,
|
||||
},
|
||||
"meta-llama/Llama-2-70b-chat-hf": {
|
||||
id: "meta-llama/Llama-2-70b-chat-hf",
|
||||
organization: "Meta",
|
||||
name: "LLaMA-2 Chat (70B)",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"meta-llama/Llama-2-13b-chat-hf": {
|
||||
id: "meta-llama/Llama-2-13b-chat-hf",
|
||||
organization: "Meta",
|
||||
name: "LLaMA-2 Chat (13B)",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"meta-llama/Llama-2-7b-chat-hf": {
|
||||
id: "meta-llama/Llama-2-7b-chat-hf",
|
||||
organization: "Meta",
|
||||
name: "LLaMA-2 Chat (7B)",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"mistralai/Mistral-7B-Instruct-v0.1": {
|
||||
id: "mistralai/Mistral-7B-Instruct-v0.1",
|
||||
organization: "mistralai",
|
||||
name: "Mistral (7B) Instruct",
|
||||
maxLength: 8192,
|
||||
},
|
||||
"mistralai/Mistral-7B-Instruct-v0.2": {
|
||||
id: "mistralai/Mistral-7B-Instruct-v0.2",
|
||||
organization: "mistralai",
|
||||
name: "Mistral (7B) Instruct v0.2",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"mistralai/Mixtral-8x7B-Instruct-v0.1": {
|
||||
id: "mistralai/Mixtral-8x7B-Instruct-v0.1",
|
||||
organization: "mistralai",
|
||||
name: "Mixtral-8x7B Instruct (46.7B)",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"NousResearch/Nous-Capybara-7B-V1p9": {
|
||||
id: "NousResearch/Nous-Capybara-7B-V1p9",
|
||||
organization: "NousResearch",
|
||||
name: "Nous Capybara v1.9 (7B)",
|
||||
maxLength: 8192,
|
||||
},
|
||||
"NousResearch/Nous-Hermes-2-Mistral-7B-DPO": {
|
||||
id: "NousResearch/Nous-Hermes-2-Mistral-7B-DPO",
|
||||
organization: "NousResearch",
|
||||
name: "Nous Hermes 2 - Mistral DPO (7B)",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO": {
|
||||
id: "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO",
|
||||
organization: "NousResearch",
|
||||
name: "Nous Hermes 2 - Mixtral 8x7B-DPO (46.7B)",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"NousResearch/Nous-Hermes-2-Mixtral-8x7B-SFT": {
|
||||
id: "NousResearch/Nous-Hermes-2-Mixtral-8x7B-SFT",
|
||||
organization: "NousResearch",
|
||||
name: "Nous Hermes 2 - Mixtral 8x7B-SFT (46.7B)",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"NousResearch/Nous-Hermes-llama-2-7b": {
|
||||
id: "NousResearch/Nous-Hermes-llama-2-7b",
|
||||
organization: "NousResearch",
|
||||
|
@ -113,66 +185,96 @@ const MODELS = {
|
|||
name: "Nous Hermes Llama-2 (13B)",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"NousResearch/Nous-Hermes-Llama2-70b": {
|
||||
id: "NousResearch/Nous-Hermes-Llama2-70b",
|
||||
organization: "NousResearch",
|
||||
name: "Nous Hermes Llama-2 (70B)",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"NousResearch/Nous-Hermes-2-Yi-34B": {
|
||||
id: "NousResearch/Nous-Hermes-2-Yi-34B",
|
||||
organization: "NousResearch",
|
||||
name: "Nous Hermes-2 Yi (34B)",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"NousResearch/Nous-Capybara-7B-V1p9": {
|
||||
id: "NousResearch/Nous-Capybara-7B-V1p9",
|
||||
organization: "NousResearch",
|
||||
name: "Nous Capybara v1.9 (7B)",
|
||||
maxLength: 8192,
|
||||
},
|
||||
"openchat/openchat-3.5-1210": {
|
||||
id: "openchat/openchat-3.5-1210",
|
||||
organization: "OpenChat",
|
||||
name: "OpenChat 3.5 1210 (7B)",
|
||||
name: "OpenChat 3.5 (7B)",
|
||||
maxLength: 8192,
|
||||
},
|
||||
"teknium/OpenHermes-2-Mistral-7B": {
|
||||
id: "teknium/OpenHermes-2-Mistral-7B",
|
||||
organization: "teknium",
|
||||
name: "OpenHermes-2-Mistral (7B)",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"teknium/OpenHermes-2p5-Mistral-7B": {
|
||||
id: "teknium/OpenHermes-2p5-Mistral-7B",
|
||||
organization: "teknium",
|
||||
name: "OpenHermes-2.5-Mistral (7B)",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"Open-Orca/Mistral-7B-OpenOrca": {
|
||||
id: "Open-Orca/Mistral-7B-OpenOrca",
|
||||
organization: "OpenOrca",
|
||||
name: "OpenOrca Mistral (7B) 8K",
|
||||
maxLength: 8192,
|
||||
},
|
||||
"garage-bAInd/Platypus2-70B-instruct": {
|
||||
id: "garage-bAInd/Platypus2-70B-instruct",
|
||||
organization: "garage-bAInd",
|
||||
name: "Platypus2 Instruct (70B)",
|
||||
maxLength: 4096,
|
||||
"Qwen/Qwen1.5-0.5B-Chat": {
|
||||
id: "Qwen/Qwen1.5-0.5B-Chat",
|
||||
organization: "Qwen",
|
||||
name: "Qwen 1.5 Chat (0.5B)",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"togethercomputer/Pythia-Chat-Base-7B-v0.16": {
|
||||
id: "togethercomputer/Pythia-Chat-Base-7B-v0.16",
|
||||
organization: "Together",
|
||||
name: "Pythia-Chat-Base (7B)",
|
||||
"Qwen/Qwen1.5-1.8B-Chat": {
|
||||
id: "Qwen/Qwen1.5-1.8B-Chat",
|
||||
organization: "Qwen",
|
||||
name: "Qwen 1.5 Chat (1.8B)",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"Qwen/Qwen1.5-4B-Chat": {
|
||||
id: "Qwen/Qwen1.5-4B-Chat",
|
||||
organization: "Qwen",
|
||||
name: "Qwen 1.5 Chat (4B)",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"Qwen/Qwen1.5-7B-Chat": {
|
||||
id: "Qwen/Qwen1.5-7B-Chat",
|
||||
organization: "Qwen",
|
||||
name: "Qwen 1.5 Chat (7B)",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"Qwen/Qwen1.5-14B-Chat": {
|
||||
id: "Qwen/Qwen1.5-14B-Chat",
|
||||
organization: "Qwen",
|
||||
name: "Qwen 1.5 Chat (14B)",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"Qwen/Qwen1.5-32B-Chat": {
|
||||
id: "Qwen/Qwen1.5-32B-Chat",
|
||||
organization: "Qwen",
|
||||
name: "Qwen 1.5 Chat (32B)",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"Qwen/Qwen1.5-72B-Chat": {
|
||||
id: "Qwen/Qwen1.5-72B-Chat",
|
||||
organization: "Qwen",
|
||||
name: "Qwen 1.5 Chat (72B)",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"snorkelai/Snorkel-Mistral-PairRM-DPO": {
|
||||
id: "snorkelai/Snorkel-Mistral-PairRM-DPO",
|
||||
organization: "Snorkel AI",
|
||||
name: "Snorkel Mistral PairRM DPO (7B)",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"togethercomputer/alpaca-7b": {
|
||||
id: "togethercomputer/alpaca-7b",
|
||||
organization: "Stanford",
|
||||
name: "Alpaca (7B)",
|
||||
maxLength: 2048,
|
||||
},
|
||||
"togethercomputer/Qwen-7B-Chat": {
|
||||
id: "togethercomputer/Qwen-7B-Chat",
|
||||
organization: "Qwen",
|
||||
name: "Qwen-Chat (7B)",
|
||||
"teknium/OpenHermes-2-Mistral-7B": {
|
||||
id: "teknium/OpenHermes-2-Mistral-7B",
|
||||
organization: "Teknium",
|
||||
name: "OpenHermes-2-Mistral (7B)",
|
||||
maxLength: 8192,
|
||||
},
|
||||
"teknium/OpenHermes-2p5-Mistral-7B": {
|
||||
id: "teknium/OpenHermes-2p5-Mistral-7B",
|
||||
organization: "Teknium",
|
||||
name: "OpenHermes-2.5-Mistral (7B)",
|
||||
maxLength: 8192,
|
||||
},
|
||||
"togethercomputer/Llama-2-7B-32K-Instruct": {
|
||||
id: "togethercomputer/Llama-2-7B-32K-Instruct",
|
||||
organization: "Together",
|
||||
name: "LLaMA-2-7B-32K-Instruct (7B)",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"togethercomputer/RedPajama-INCITE-Chat-3B-v1": {
|
||||
id: "togethercomputer/RedPajama-INCITE-Chat-3B-v1",
|
||||
organization: "Together",
|
||||
|
@ -185,40 +287,34 @@ const MODELS = {
|
|||
name: "RedPajama-INCITE Chat (7B)",
|
||||
maxLength: 2048,
|
||||
},
|
||||
"upstage/SOLAR-0-70b-16bit": {
|
||||
id: "upstage/SOLAR-0-70b-16bit",
|
||||
organization: "Upstage",
|
||||
name: "SOLAR v0 (70B)",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"togethercomputer/StripedHyena-Nous-7B": {
|
||||
id: "togethercomputer/StripedHyena-Nous-7B",
|
||||
organization: "Together",
|
||||
name: "StripedHyena Nous (7B)",
|
||||
maxLength: 32768,
|
||||
},
|
||||
"lmsys/vicuna-7b-v1.5": {
|
||||
id: "lmsys/vicuna-7b-v1.5",
|
||||
organization: "LM Sys",
|
||||
name: "Vicuna v1.5 (7B)",
|
||||
"Undi95/ReMM-SLERP-L2-13B": {
|
||||
id: "Undi95/ReMM-SLERP-L2-13B",
|
||||
organization: "Undi95",
|
||||
name: "ReMM SLERP L2 (13B)",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"lmsys/vicuna-13b-v1.5": {
|
||||
id: "lmsys/vicuna-13b-v1.5",
|
||||
organization: "LM Sys",
|
||||
name: "Vicuna v1.5 (13B)",
|
||||
"Undi95/Toppy-M-7B": {
|
||||
id: "Undi95/Toppy-M-7B",
|
||||
organization: "Undi95",
|
||||
name: "Toppy M (7B)",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"lmsys/vicuna-13b-v1.5-16k": {
|
||||
id: "lmsys/vicuna-13b-v1.5-16k",
|
||||
organization: "LM Sys",
|
||||
name: "Vicuna v1.5 16K (13B)",
|
||||
maxLength: 16384,
|
||||
"WizardLM/WizardLM-13B-V1.2": {
|
||||
id: "WizardLM/WizardLM-13B-V1.2",
|
||||
organization: "WizardLM",
|
||||
name: "WizardLM v1.2 (13B)",
|
||||
maxLength: 4096,
|
||||
},
|
||||
"zero-one-ai/Yi-34B-Chat": {
|
||||
id: "zero-one-ai/Yi-34B-Chat",
|
||||
organization: "01.AI",
|
||||
name: "01-ai Yi Chat (34B)",
|
||||
"upstage/SOLAR-10.7B-Instruct-v1.0": {
|
||||
id: "upstage/SOLAR-10.7B-Instruct-v1.0",
|
||||
organization: "upstage",
|
||||
name: "Upstage SOLAR Instruct v1 (11B)",
|
||||
maxLength: 4096,
|
||||
},
|
||||
};
|
||||
|
|
|
@ -1,39 +1,55 @@
|
|||
| Organization | Model Name | Model String for API | Max Seq Length |
|
||||
| ------------- | ---------------------------- | -------------------------------------------- | -------------- |
|
||||
| Stanford | Alpaca (7B) | togethercomputer/alpaca-7b | 2048 |
|
||||
| Austism | Chronos Hermes (13B) | Austism/chronos-hermes-13b | 2048 |
|
||||
| Meta | Code Llama Instruct (13B) | togethercomputer/CodeLlama-13b-Instruct | 8192 |
|
||||
| Meta | Code Llama Instruct (34B) | togethercomputer/CodeLlama-34b-Instruct | 8192 |
|
||||
| Meta | Code Llama Instruct (7B) | togethercomputer/CodeLlama-7b-Instruct | 8192 |
|
||||
| DiscoResearch | DiscoLM Mixtral 8x7b | DiscoResearch/DiscoLM-mixtral-8x7b-v2 | 32768 |
|
||||
| TII UAE | Falcon Instruct (40B) | togethercomputer/falcon-40b-instruct | 2048 |
|
||||
| TII UAE | Falcon Instruct (7B) | togethercomputer/falcon-7b-instruct | 2048 |
|
||||
| Together | GPT-NeoXT-Chat-Base (20B) | togethercomputer/GPT-NeoXT-Chat-Base-20B | 2048 |
|
||||
| Meta | LLaMA-2 Chat (13B) | togethercomputer/llama-2-13b-chat | 4096 |
|
||||
| Meta | LLaMA-2 Chat (70B) | togethercomputer/llama-2-70b-chat | 4096 |
|
||||
| Meta | LLaMA-2 Chat (7B) | togethercomputer/llama-2-7b-chat | 4096 |
|
||||
| Together | LLaMA-2-7B-32K-Instruct (7B) | togethercomputer/Llama-2-7B-32K-Instruct | 32768 |
|
||||
| MistralAI | Mistral (7B) Instruct v0.1 | mistralai/Mistral-7B-Instruct-v0.1 | 4096 |
|
||||
| MistralAI | Mistral (7B) Instruct v0.2 | mistralai/Mistral-7B-Instruct-v0.2 | 32768 |
|
||||
| MistralAI | Mixtral-8x7B Instruct | mistralai/Mixtral-8x7B-Instruct-v0.1 | 32768 |
|
||||
| Gryphe | MythoMax-L2 (13B) | Gryphe/MythoMax-L2-13b | 4096 |
|
||||
| NousResearch | Nous Hermes LLaMA-2 (7B) | NousResearch/Nous-Hermes-llama-2-7b | 4096 |
|
||||
| NousResearch | Nous Hermes Llama-2 (13B) | NousResearch/Nous-Hermes-Llama2-13b | 4096 |
|
||||
| NousResearch | Nous Hermes Llama-2 (70B) | NousResearch/Nous-Hermes-Llama2-70b | 4096 |
|
||||
| NousResearch | Nous Hermes-2 Yi (34B) | NousResearch/Nous-Hermes-2-Yi-34B | 4096 |
|
||||
| NousResearch | Nous Capybara v1.9 (7B) | NousResearch/Nous-Capybara-7B-V1p9 | 8192 |
|
||||
| OpenChat | OpenChat 3.5 1210 (7B) | openchat/openchat-3.5-1210 | 8192 |
|
||||
| teknium | OpenHermes-2-Mistral (7B) | teknium/OpenHermes-2-Mistral-7B | 4096 |
|
||||
| teknium | OpenHermes-2.5-Mistral (7B) | teknium/OpenHermes-2p5-Mistral-7B | 4096 |
|
||||
| OpenOrca | OpenOrca Mistral (7B) 8K | Open-Orca/Mistral-7B-OpenOrca | 8192 |
|
||||
| garage-bAInd | Platypus2 Instruct (70B) | garage-bAInd/Platypus2-70B-instruct | 4096 |
|
||||
| Together | Pythia-Chat-Base (7B) | togethercomputer/Pythia-Chat-Base-7B-v0.16 | 2048 |
|
||||
| Qwen | Qwen-Chat (7B) | togethercomputer/Qwen-7B-Chat | 8192 |
|
||||
| Together | RedPajama-INCITE Chat (3B) | togethercomputer/RedPajama-INCITE-Chat-3B-v1 | 2048 |
|
||||
| Together | RedPajama-INCITE Chat (7B) | togethercomputer/RedPajama-INCITE-7B-Chat | 2048 |
|
||||
| Upstage | SOLAR v0 (70B) | upstage/SOLAR-0-70b-16bit | 4096 |
|
||||
| Together | StripedHyena Nous (7B) | togethercomputer/StripedHyena-Nous-7B | 32768 |
|
||||
| LM Sys | Vicuna v1.5 (7B) | lmsys/vicuna-7b-v1.5 | 4096 |
|
||||
| LM Sys | Vicuna v1.5 (13B) | lmsys/vicuna-13b-v1.5 | 4096 |
|
||||
| LM Sys | Vicuna v1.5 16K (13B) | lmsys/vicuna-13b-v1.5-16k | 16384 |
|
||||
| 01.AI | 01-ai Yi Chat (34B) | zero-one-ai/Yi-34B-Chat | 4096 |
|
||||
| Organization | Model Name | Model String for API | Context length |
|
||||
| --- | --- | --- | --- |
|
||||
| 01.AI | 01-ai Yi Chat (34B) | zero-one-ai/Yi-34B-Chat | 4096 |
|
||||
| Allen AI | OLMo Instruct (7B) | allenai/OLMo-7B-Instruct | 2048 |
|
||||
| Allen AI | OLMo Twin-2T (7B) | allenai/OLMo-7B-Twin-2T | 2048 |
|
||||
| Allen AI | OLMo (7B) | allenai/OLMo-7B | 2048 |
|
||||
| Austism | Chronos Hermes (13B) | Austism/chronos-hermes-13b | 2048 |
|
||||
| cognitivecomputations | Dolphin 2.5 Mixtral 8x7b | cognitivecomputations/dolphin-2.5-mixtral-8x7b | 32768 |
|
||||
| databricks | DBRX Instruct | databricks/dbrx-instruct | 32000 |
|
||||
| DeepSeek | Deepseek Coder Instruct (33B) | deepseek-ai/deepseek-coder-33b-instruct | 16384 |
|
||||
| DeepSeek | DeepSeek LLM Chat (67B) | deepseek-ai/deepseek-llm-67b-chat | 4096 |
|
||||
| garage-bAInd | Platypus2 Instruct (70B) | garage-bAInd/Platypus2-70B-instruct | 4096 |
|
||||
| Google | Gemma Instruct (2B) | google/gemma-2b-it | 8192 |
|
||||
| Google | Gemma Instruct (7B) | google/gemma-7b-it | 8192 |
|
||||
| Gryphe | MythoMax-L2 (13B) | Gryphe/MythoMax-L2-13b | 4096 |
|
||||
| LM Sys | Vicuna v1.5 (13B) | lmsys/vicuna-13b-v1.5 | 4096 |
|
||||
| LM Sys | Vicuna v1.5 (7B) | lmsys/vicuna-7b-v1.5 | 4096 |
|
||||
| Meta | Code Llama Instruct (13B) | codellama/CodeLlama-13b-Instruct-hf | 16384 |
|
||||
| Meta | Code Llama Instruct (34B) | codellama/CodeLlama-34b-Instruct-hf | 16384 |
|
||||
| Meta | Code Llama Instruct (70B) | codellama/CodeLlama-70b-Instruct-hf | 4096 |
|
||||
| Meta | Code Llama Instruct (7B) | codellama/CodeLlama-7b-Instruct-hf | 16384 |
|
||||
| Meta | LLaMA-2 Chat (70B) | meta-llama/Llama-2-70b-chat-hf | 4096 |
|
||||
| Meta | LLaMA-2 Chat (13B) | meta-llama/Llama-2-13b-chat-hf | 4096 |
|
||||
| Meta | LLaMA-2 Chat (7B) | meta-llama/Llama-2-7b-chat-hf | 4096 |
|
||||
| mistralai | Mistral (7B) Instruct | mistralai/Mistral-7B-Instruct-v0.1 | 8192 |
|
||||
| mistralai | Mistral (7B) Instruct v0.2 | mistralai/Mistral-7B-Instruct-v0.2 | 32768 |
|
||||
| mistralai | Mixtral-8x7B Instruct (46.7B) | mistralai/Mixtral-8x7B-Instruct-v0.1 | 32768 |
|
||||
| NousResearch | Nous Capybara v1.9 (7B) | NousResearch/Nous-Capybara-7B-V1p9 | 8192 |
|
||||
| NousResearch | Nous Hermes 2 - Mistral DPO (7B) | NousResearch/Nous-Hermes-2-Mistral-7B-DPO | 32768 |
|
||||
| NousResearch | Nous Hermes 2 - Mixtral 8x7B-DPO (46.7B) | NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO | 32768 |
|
||||
| NousResearch | Nous Hermes 2 - Mixtral 8x7B-SFT (46.7B) | NousResearch/Nous-Hermes-2-Mixtral-8x7B-SFT | 32768 |
|
||||
| NousResearch | Nous Hermes LLaMA-2 (7B) | NousResearch/Nous-Hermes-llama-2-7b | 4096 |
|
||||
| NousResearch | Nous Hermes Llama-2 (13B) | NousResearch/Nous-Hermes-Llama2-13b | 4096 |
|
||||
| NousResearch | Nous Hermes-2 Yi (34B) | NousResearch/Nous-Hermes-2-Yi-34B | 4096 |
|
||||
| OpenChat | OpenChat 3.5 (7B) | openchat/openchat-3.5-1210 | 8192 |
|
||||
| OpenOrca | OpenOrca Mistral (7B) 8K | Open-Orca/Mistral-7B-OpenOrca | 8192 |
|
||||
| Qwen | Qwen 1.5 Chat (0.5B) | Qwen/Qwen1.5-0.5B-Chat | 32768 |
|
||||
| Qwen | Qwen 1.5 Chat (1.8B) | Qwen/Qwen1.5-1.8B-Chat | 32768 |
|
||||
| Qwen | Qwen 1.5 Chat (4B) | Qwen/Qwen1.5-4B-Chat | 32768 |
|
||||
| Qwen | Qwen 1.5 Chat (7B) | Qwen/Qwen1.5-7B-Chat | 32768 |
|
||||
| Qwen | Qwen 1.5 Chat (14B) | Qwen/Qwen1.5-14B-Chat | 32768 |
|
||||
| Qwen | Qwen 1.5 Chat (32B) | Qwen/Qwen1.5-32B-Chat | 32768 |
|
||||
| Qwen | Qwen 1.5 Chat (72B) | Qwen/Qwen1.5-72B-Chat | 32768 |
|
||||
| Snorkel AI | Snorkel Mistral PairRM DPO (7B) | snorkelai/Snorkel-Mistral-PairRM-DPO | 32768 |
|
||||
| Stanford | Alpaca (7B) | togethercomputer/alpaca-7b | 2048 |
|
||||
| Teknium | OpenHermes-2-Mistral (7B) | teknium/OpenHermes-2-Mistral-7B | 8192 |
|
||||
| Teknium | OpenHermes-2.5-Mistral (7B) | teknium/OpenHermes-2p5-Mistral-7B | 8192 |
|
||||
| Together | LLaMA-2-7B-32K-Instruct (7B) | togethercomputer/Llama-2-7B-32K-Instruct | 32768 |
|
||||
| Together | RedPajama-INCITE Chat (3B) | togethercomputer/RedPajama-INCITE-Chat-3B-v1 | 2048 |
|
||||
| Together | RedPajama-INCITE Chat (7B) | togethercomputer/RedPajama-INCITE-7B-Chat | 2048 |
|
||||
| Together | StripedHyena Nous (7B) | togethercomputer/StripedHyena-Nous-7B | 32768 |
|
||||
| Undi95 | ReMM SLERP L2 (13B) | Undi95/ReMM-SLERP-L2-13B | 4096 |
|
||||
| Undi95 | Toppy M (7B) | Undi95/Toppy-M-7B | 4096 |
|
||||
| WizardLM | WizardLM v1.2 (13B) | WizardLM/WizardLM-13B-V1.2 | 4096 |
|
||||
| upstage | Upstage SOLAR Instruct v1 (11B) | upstage/SOLAR-10.7B-Instruct-v1.0 | 4096 |
|
|
@ -8,7 +8,9 @@
|
|||
// copy outputs into the export in ../models.js
|
||||
|
||||
// Update the date below if you run this again because TogetherAI added new models.
|
||||
// Last Collected: Jan 10, 2023
|
||||
// Last Collected: Apr 14, 2024
|
||||
// Since last collection Together's docs are broken. I just copied the HTML table
|
||||
// and had claude3 convert to markdown and it works well enough.
|
||||
|
||||
import fs from "fs";
|
||||
|
||||
|
|
Loading…
Add table
Reference in a new issue