Bump all static model providers ()

This commit is contained in:
Timothy Carambat 2024-04-14 12:55:21 -07:00 committed by GitHub
parent 6fde5570b3
commit 8306098b08
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
8 changed files with 573 additions and 339 deletions
server/utils/AiProviders

View file

@ -1,10 +1,4 @@
const MODELS = {
"nousresearch/nous-capybara-34b": {
id: "nousresearch/nous-capybara-34b",
name: "Nous: Capybara 34B",
organization: "Nousresearch",
maxLength: 32768,
},
"openrouter/auto": {
id: "openrouter/auto",
name: "Auto (best for prompt)",
@ -21,6 +15,12 @@ const MODELS = {
id: "mistralai/mistral-7b-instruct:free",
name: "Mistral 7B Instruct (free)",
organization: "Mistralai",
maxLength: 32768,
},
"openchat/openchat-7b:free": {
id: "openchat/openchat-7b:free",
name: "OpenChat 3.5 (free)",
organization: "Openchat",
maxLength: 8192,
},
"gryphe/mythomist-7b:free": {
@ -45,13 +45,7 @@ const MODELS = {
id: "google/gemma-7b-it:free",
name: "Google: Gemma 7B (free)",
organization: "Google",
maxLength: 8000,
},
"jondurbin/bagel-34b": {
id: "jondurbin/bagel-34b",
name: "Bagel 34B v0.2",
organization: "Jondurbin",
maxLength: 8000,
maxLength: 8192,
},
"jebcarter/psyfighter-13b": {
id: "jebcarter/psyfighter-13b",
@ -65,54 +59,12 @@ const MODELS = {
organization: "Koboldai",
maxLength: 4096,
},
"neversleep/noromaid-mixtral-8x7b-instruct": {
id: "neversleep/noromaid-mixtral-8x7b-instruct",
name: "Noromaid Mixtral 8x7B Instruct",
organization: "Neversleep",
maxLength: 8000,
},
"nousresearch/nous-hermes-llama2-13b": {
id: "nousresearch/nous-hermes-llama2-13b",
name: "Nous: Hermes 13B",
organization: "Nousresearch",
maxLength: 4096,
},
"meta-llama/codellama-34b-instruct": {
id: "meta-llama/codellama-34b-instruct",
name: "Meta: CodeLlama 34B Instruct",
organization: "Meta-llama",
maxLength: 8192,
},
"phind/phind-codellama-34b": {
id: "phind/phind-codellama-34b",
name: "Phind: CodeLlama 34B v2",
organization: "Phind",
maxLength: 4096,
},
"intel/neural-chat-7b": {
id: "intel/neural-chat-7b",
name: "Neural Chat 7B v3.1",
organization: "Intel",
maxLength: 4096,
},
"mistralai/mixtral-8x7b-instruct": {
id: "mistralai/mixtral-8x7b-instruct",
name: "Mistral: Mixtral 8x7B Instruct",
organization: "Mistralai",
maxLength: 32768,
},
"nousresearch/nous-hermes-2-mixtral-8x7b-dpo": {
id: "nousresearch/nous-hermes-2-mixtral-8x7b-dpo",
name: "Nous: Hermes 2 Mixtral 8x7B DPO",
organization: "Nousresearch",
maxLength: 32000,
},
"nousresearch/nous-hermes-2-mixtral-8x7b-sft": {
id: "nousresearch/nous-hermes-2-mixtral-8x7b-sft",
name: "Nous: Hermes 2 Mixtral 8x7B SFT",
organization: "Nousresearch",
maxLength: 32000,
},
"haotian-liu/llava-13b": {
id: "haotian-liu/llava-13b",
name: "Llava 13B",
@ -143,30 +95,12 @@ const MODELS = {
organization: "Pygmalionai",
maxLength: 8192,
},
"undi95/remm-slerp-l2-13b-6k": {
id: "undi95/remm-slerp-l2-13b-6k",
name: "ReMM SLERP 13B 6k",
organization: "Undi95",
maxLength: 6144,
},
"gryphe/mythomax-l2-13b": {
id: "gryphe/mythomax-l2-13b",
name: "MythoMax 13B",
organization: "Gryphe",
maxLength: 4096,
},
"xwin-lm/xwin-lm-70b": {
id: "xwin-lm/xwin-lm-70b",
name: "Xwin 70B",
organization: "Xwin-lm",
maxLength: 8192,
},
"gryphe/mythomax-l2-13b-8k": {
id: "gryphe/mythomax-l2-13b-8k",
name: "MythoMax 13B 8k",
organization: "Gryphe",
maxLength: 8192,
},
"alpindale/goliath-120b": {
id: "alpindale/goliath-120b",
name: "Goliath 120B",
@ -185,15 +119,27 @@ const MODELS = {
organization: "Gryphe",
maxLength: 32768,
},
"sophosympatheia/midnight-rose-70b": {
id: "sophosympatheia/midnight-rose-70b",
name: "Midnight Rose 70B",
organization: "Sophosympatheia",
maxLength: 4096,
},
"undi95/remm-slerp-l2-13b:extended": {
id: "undi95/remm-slerp-l2-13b:extended",
name: "ReMM SLERP 13B (extended)",
organization: "Undi95",
maxLength: 6144,
},
"mancer/weaver": {
id: "mancer/weaver",
name: "Mancer: Weaver (alpha)",
organization: "Mancer",
maxLength: 8000,
},
"nousresearch/nous-hermes-llama2-70b": {
id: "nousresearch/nous-hermes-llama2-70b",
name: "Nous: Hermes 70B",
"nousresearch/nous-hermes-llama2-13b": {
id: "nousresearch/nous-hermes-llama2-13b",
name: "Nous: Hermes 13B",
organization: "Nousresearch",
maxLength: 4096,
},
@ -203,12 +149,24 @@ const MODELS = {
organization: "Nousresearch",
maxLength: 4096,
},
"meta-llama/codellama-34b-instruct": {
id: "meta-llama/codellama-34b-instruct",
name: "Meta: CodeLlama 34B Instruct",
organization: "Meta-llama",
maxLength: 8192,
},
"codellama/codellama-70b-instruct": {
id: "codellama/codellama-70b-instruct",
name: "Meta: CodeLlama 70B Instruct",
organization: "Codellama",
maxLength: 2048,
},
"phind/phind-codellama-34b": {
id: "phind/phind-codellama-34b",
name: "Phind: CodeLlama 34B v2",
organization: "Phind",
maxLength: 4096,
},
"teknium/openhermes-2-mistral-7b": {
id: "teknium/openhermes-2-mistral-7b",
name: "OpenHermes 2 Mistral 7B",
@ -227,12 +185,6 @@ const MODELS = {
organization: "Undi95",
maxLength: 4096,
},
"undi95/toppy-m-7b": {
id: "undi95/toppy-m-7b",
name: "Toppy M 7B",
organization: "Undi95",
maxLength: 4096,
},
"openrouter/cinematika-7b": {
id: "openrouter/cinematika-7b",
name: "Cinematika 7B (alpha)",
@ -271,7 +223,7 @@ const MODELS = {
},
"mistralai/mixtral-8x7b": {
id: "mistralai/mixtral-8x7b",
name: "Mistral: Mixtral 8x7B (base)",
name: "Mixtral 8x7B (base)",
organization: "Mistralai",
maxLength: 32768,
},
@ -281,6 +233,12 @@ const MODELS = {
organization: "Nousresearch",
maxLength: 4096,
},
"nousresearch/nous-hermes-2-mixtral-8x7b-sft": {
id: "nousresearch/nous-hermes-2-mixtral-8x7b-sft",
name: "Nous: Hermes 2 Mixtral 8x7B SFT",
organization: "Nousresearch",
maxLength: 32000,
},
"nousresearch/nous-hermes-2-mistral-7b-dpo": {
id: "nousresearch/nous-hermes-2-mistral-7b-dpo",
name: "Nous: Hermes 2 Mistral 7B DPO",
@ -303,7 +261,7 @@ const MODELS = {
id: "openai/gpt-3.5-turbo",
name: "OpenAI: GPT-3.5 Turbo",
organization: "Openai",
maxLength: 4095,
maxLength: 16385,
},
"openai/gpt-3.5-turbo-0125": {
id: "openai/gpt-3.5-turbo-0125",
@ -335,9 +293,15 @@ const MODELS = {
organization: "Openai",
maxLength: 16385,
},
"openai/gpt-4-turbo": {
id: "openai/gpt-4-turbo",
name: "OpenAI: GPT-4 Turbo",
organization: "Openai",
maxLength: 128000,
},
"openai/gpt-4-turbo-preview": {
id: "openai/gpt-4-turbo-preview",
name: "OpenAI: GPT-4 Turbo (preview)",
name: "OpenAI: GPT-4 Turbo Preview",
organization: "Openai",
maxLength: 128000,
},
@ -373,7 +337,7 @@ const MODELS = {
},
"openai/gpt-4-vision-preview": {
id: "openai/gpt-4-vision-preview",
name: "OpenAI: GPT-4 Vision (preview)",
name: "OpenAI: GPT-4 Vision",
organization: "Openai",
maxLength: 128000,
},
@ -387,37 +351,43 @@ const MODELS = {
id: "google/palm-2-chat-bison",
name: "Google: PaLM 2 Chat",
organization: "Google",
maxLength: 36864,
maxLength: 25804,
},
"google/palm-2-codechat-bison": {
id: "google/palm-2-codechat-bison",
name: "Google: PaLM 2 Code Chat",
organization: "Google",
maxLength: 28672,
maxLength: 20070,
},
"google/palm-2-chat-bison-32k": {
id: "google/palm-2-chat-bison-32k",
name: "Google: PaLM 2 Chat 32k",
organization: "Google",
maxLength: 131072,
maxLength: 91750,
},
"google/palm-2-codechat-bison-32k": {
id: "google/palm-2-codechat-bison-32k",
name: "Google: PaLM 2 Code Chat 32k",
organization: "Google",
maxLength: 131072,
maxLength: 91750,
},
"google/gemini-pro": {
id: "google/gemini-pro",
name: "Google: Gemini Pro (preview)",
name: "Google: Gemini Pro 1.0",
organization: "Google",
maxLength: 131040,
maxLength: 91728,
},
"google/gemini-pro-vision": {
id: "google/gemini-pro-vision",
name: "Google: Gemini Pro Vision (preview)",
name: "Google: Gemini Pro Vision 1.0",
organization: "Google",
maxLength: 65536,
maxLength: 45875,
},
"google/gemini-pro-1.5": {
id: "google/gemini-pro-1.5",
name: "Google: Gemini Pro 1.5 (preview)",
organization: "Google",
maxLength: 2800000,
},
"perplexity/pplx-70b-online": {
id: "perplexity/pplx-70b-online",
@ -443,18 +413,96 @@ const MODELS = {
organization: "Perplexity",
maxLength: 4096,
},
"perplexity/sonar-small-chat": {
id: "perplexity/sonar-small-chat",
name: "Perplexity: Sonar 7B",
organization: "Perplexity",
maxLength: 16384,
},
"perplexity/sonar-medium-chat": {
id: "perplexity/sonar-medium-chat",
name: "Perplexity: Sonar 8x7B",
organization: "Perplexity",
maxLength: 16384,
},
"perplexity/sonar-small-online": {
id: "perplexity/sonar-small-online",
name: "Perplexity: Sonar 7B Online",
organization: "Perplexity",
maxLength: 12000,
},
"perplexity/sonar-medium-online": {
id: "perplexity/sonar-medium-online",
name: "Perplexity: Sonar 8x7B Online",
organization: "Perplexity",
maxLength: 12000,
},
"fireworks/mixtral-8x22b-instruct-preview": {
id: "fireworks/mixtral-8x22b-instruct-preview",
name: "Fireworks Mixtral 8x22B Instruct OH (preview)",
organization: "Fireworks",
maxLength: 8192,
},
"anthropic/claude-3-opus": {
id: "anthropic/claude-3-opus",
name: "Anthropic: Claude 3 Opus",
organization: "Anthropic",
maxLength: 200000,
},
"anthropic/claude-3-sonnet": {
id: "anthropic/claude-3-sonnet",
name: "Anthropic: Claude 3 Sonnet",
organization: "Anthropic",
maxLength: 200000,
},
"anthropic/claude-3-haiku": {
id: "anthropic/claude-3-haiku",
name: "Anthropic: Claude 3 Haiku",
organization: "Anthropic",
maxLength: 200000,
},
"anthropic/claude-3-opus:beta": {
id: "anthropic/claude-3-opus:beta",
name: "Anthropic: Claude 3 Opus (self-moderated)",
organization: "Anthropic",
maxLength: 200000,
},
"anthropic/claude-3-sonnet:beta": {
id: "anthropic/claude-3-sonnet:beta",
name: "Anthropic: Claude 3 Sonnet (self-moderated)",
organization: "Anthropic",
maxLength: 200000,
},
"anthropic/claude-3-haiku:beta": {
id: "anthropic/claude-3-haiku:beta",
name: "Anthropic: Claude 3 Haiku (self-moderated)",
organization: "Anthropic",
maxLength: 200000,
},
"meta-llama/llama-2-70b-chat": {
id: "meta-llama/llama-2-70b-chat",
name: "Meta: Llama v2 70B Chat",
organization: "Meta-llama",
maxLength: 4096,
},
"nousresearch/nous-capybara-34b": {
id: "nousresearch/nous-capybara-34b",
name: "Nous: Capybara 34B",
organization: "Nousresearch",
maxLength: 32768,
},
"jondurbin/airoboros-l2-70b": {
id: "jondurbin/airoboros-l2-70b",
name: "Airoboros 70B",
organization: "Jondurbin",
maxLength: 4096,
},
"jondurbin/bagel-34b": {
id: "jondurbin/bagel-34b",
name: "Bagel 34B v0.2",
organization: "Jondurbin",
maxLength: 8000,
},
"austism/chronos-hermes-13b": {
id: "austism/chronos-hermes-13b",
name: "Chronos Hermes 13B v2",
@ -465,7 +513,13 @@ const MODELS = {
id: "mistralai/mistral-7b-instruct",
name: "Mistral 7B Instruct",
organization: "Mistralai",
maxLength: 8192,
maxLength: 32768,
},
"gryphe/mythomax-l2-13b": {
id: "gryphe/mythomax-l2-13b",
name: "MythoMax 13B",
organization: "Gryphe",
maxLength: 4096,
},
"openchat/openchat-7b": {
id: "openchat/openchat-7b",
@ -473,18 +527,42 @@ const MODELS = {
organization: "Openchat",
maxLength: 8192,
},
"undi95/toppy-m-7b": {
id: "undi95/toppy-m-7b",
name: "Toppy M 7B",
organization: "Undi95",
maxLength: 4096,
},
"lizpreciatior/lzlv-70b-fp16-hf": {
id: "lizpreciatior/lzlv-70b-fp16-hf",
name: "lzlv 70B",
organization: "Lizpreciatior",
maxLength: 4096,
},
"mistralai/mixtral-8x7b-instruct": {
id: "mistralai/mixtral-8x7b-instruct",
name: "Mixtral 8x7B Instruct",
organization: "Mistralai",
maxLength: 32768,
},
"cognitivecomputations/dolphin-mixtral-8x7b": {
id: "cognitivecomputations/dolphin-mixtral-8x7b",
name: "Dolphin 2.6 Mixtral 8x7B 🐬",
organization: "Cognitivecomputations",
maxLength: 32000,
},
"neversleep/noromaid-mixtral-8x7b-instruct": {
id: "neversleep/noromaid-mixtral-8x7b-instruct",
name: "Noromaid Mixtral 8x7B Instruct",
organization: "Neversleep",
maxLength: 8000,
},
"nousresearch/nous-hermes-2-mixtral-8x7b-dpo": {
id: "nousresearch/nous-hermes-2-mixtral-8x7b-dpo",
name: "Nous: Hermes 2 Mixtral 8x7B DPO",
organization: "Nousresearch",
maxLength: 32000,
},
"rwkv/rwkv-5-world-3b": {
id: "rwkv/rwkv-5-world-3b",
name: "RWKV v5 World 3B",
@ -507,7 +585,19 @@ const MODELS = {
id: "google/gemma-7b-it",
name: "Google: Gemma 7B",
organization: "Google",
maxLength: 8000,
maxLength: 8192,
},
"databricks/dbrx-instruct": {
id: "databricks/dbrx-instruct",
name: "Databricks: DBRX 132B Instruct",
organization: "Databricks",
maxLength: 32768,
},
"huggingfaceh4/zephyr-orpo-141b-a35b": {
id: "huggingfaceh4/zephyr-orpo-141b-a35b",
name: "Zephyr 141B-A35B",
organization: "Huggingfaceh4",
maxLength: 65536,
},
"anthropic/claude-2": {
id: "anthropic/claude-2",
@ -565,58 +655,124 @@ const MODELS = {
},
"anthropic/claude-2:beta": {
id: "anthropic/claude-2:beta",
name: "Anthropic: Claude v2 (experimental)",
name: "Anthropic: Claude v2 (self-moderated)",
organization: "Anthropic",
maxLength: 200000,
},
"anthropic/claude-2.1:beta": {
id: "anthropic/claude-2.1:beta",
name: "Anthropic: Claude v2.1 (experimental)",
name: "Anthropic: Claude v2.1 (self-moderated)",
organization: "Anthropic",
maxLength: 200000,
},
"anthropic/claude-2.0:beta": {
id: "anthropic/claude-2.0:beta",
name: "Anthropic: Claude v2.0 (experimental)",
name: "Anthropic: Claude v2.0 (self-moderated)",
organization: "Anthropic",
maxLength: 100000,
},
"anthropic/claude-instant-1:beta": {
id: "anthropic/claude-instant-1:beta",
name: "Anthropic: Claude Instant v1 (experimental)",
name: "Anthropic: Claude Instant v1 (self-moderated)",
organization: "Anthropic",
maxLength: 100000,
},
"mistralai/mixtral-8x22b": {
id: "mistralai/mixtral-8x22b",
name: "Mistral: Mixtral 8x22B (base)",
organization: "Mistralai",
maxLength: 65536,
},
"huggingfaceh4/zephyr-7b-beta:free": {
id: "huggingfaceh4/zephyr-7b-beta:free",
name: "Hugging Face: Zephyr 7B (free)",
organization: "Huggingfaceh4",
maxLength: 4096,
},
"openchat/openchat-7b:free": {
id: "openchat/openchat-7b:free",
name: "OpenChat 3.5 (free)",
organization: "Openchat",
"mistralai/mixtral-8x7b-instruct:nitro": {
id: "mistralai/mixtral-8x7b-instruct:nitro",
name: "Mixtral 8x7B Instruct (nitro)",
organization: "Mistralai",
maxLength: 32768,
},
"meta-llama/llama-2-70b-chat:nitro": {
id: "meta-llama/llama-2-70b-chat:nitro",
name: "Meta: Llama v2 70B Chat (nitro)",
organization: "Meta-llama",
maxLength: 4096,
},
"gryphe/mythomax-l2-13b:nitro": {
id: "gryphe/mythomax-l2-13b:nitro",
name: "MythoMax 13B (nitro)",
organization: "Gryphe",
maxLength: 4096,
},
"mistralai/mistral-7b-instruct:nitro": {
id: "mistralai/mistral-7b-instruct:nitro",
name: "Mistral 7B Instruct (nitro)",
organization: "Mistralai",
maxLength: 32768,
},
"google/gemma-7b-it:nitro": {
id: "google/gemma-7b-it:nitro",
name: "Google: Gemma 7B (nitro)",
organization: "Google",
maxLength: 8192,
},
"databricks/dbrx-instruct:nitro": {
id: "databricks/dbrx-instruct:nitro",
name: "Databricks: DBRX 132B Instruct (nitro)",
organization: "Databricks",
maxLength: 32768,
},
"gryphe/mythomax-l2-13b:extended": {
id: "gryphe/mythomax-l2-13b:extended",
name: "MythoMax 13B (extended)",
organization: "Gryphe",
maxLength: 8192,
},
"mistralai/mistral-tiny": {
id: "mistralai/mistral-tiny",
name: "Mistral: Tiny",
name: "Mistral Tiny",
organization: "Mistralai",
maxLength: 32000,
},
"mistralai/mistral-small": {
id: "mistralai/mistral-small",
name: "Mistral: Small",
name: "Mistral Small",
organization: "Mistralai",
maxLength: 32000,
},
"mistralai/mistral-medium": {
id: "mistralai/mistral-medium",
name: "Mistral: Medium",
name: "Mistral Medium",
organization: "Mistralai",
maxLength: 32000,
},
"mistralai/mistral-large": {
id: "mistralai/mistral-large",
name: "Mistral Large",
organization: "Mistralai",
maxLength: 32000,
},
"cohere/command": {
id: "cohere/command",
name: "Cohere: Command",
organization: "Cohere",
maxLength: 4096,
},
"cohere/command-r": {
id: "cohere/command-r",
name: "Cohere: Command R",
organization: "Cohere",
maxLength: 128000,
},
"cohere/command-r-plus": {
id: "cohere/command-r-plus",
name: "Cohere: Command R+",
organization: "Cohere",
maxLength: 128000,
},
};
module.exports.MODELS = MODELS;

View file

@ -6,7 +6,7 @@
// copy outputs into the export in ../models.js
// Update the date below if you run this again because OpenRouter added new models.
// Last Collected: Feb 23, 2024
// Last Collected: Apr 14, 2024
import fs from "fs";

View file

@ -19,21 +19,11 @@ const MODELS = {
name: "sonar-medium-online",
maxLength: 12000,
},
"codellama-34b-instruct": {
id: "codellama-34b-instruct",
name: "codellama-34b-instruct",
maxLength: 16384,
},
"codellama-70b-instruct": {
id: "codellama-70b-instruct",
name: "codellama-70b-instruct",
maxLength: 16384,
},
"llama-2-70b-chat": {
id: "llama-2-70b-chat",
name: "llama-2-70b-chat",
maxLength: 4096,
},
"mistral-7b-instruct": {
id: "mistral-7b-instruct",
name: "mistral-7b-instruct",
@ -44,26 +34,6 @@ const MODELS = {
name: "mixtral-8x7b-instruct",
maxLength: 16384,
},
"pplx-7b-chat": {
id: "pplx-7b-chat",
name: "pplx-7b-chat",
maxLength: 16384,
},
"pplx-7b-online": {
id: "pplx-7b-online",
name: "pplx-7b-online",
maxLength: 12000,
},
"pplx-70b-chat": {
id: "pplx-70b-chat",
name: "pplx-70b-chat",
maxLength: 8192,
},
"pplx-70b-online": {
id: "pplx-70b-online",
name: "pplx-70b-online",
maxLength: 4000,
},
};
module.exports.MODELS = MODELS;

View file

@ -1,15 +1,9 @@
| Model | Parameter Count | Context Length | Model Type |
| :-------------------------- | :-------------- | :------------- | :-------------- |
| `sonar-small-chat` | 7B | 16384 | Chat Completion |
| `sonar-small-online` | 7B | 12000 | Chat Completion |
| `sonar-medium-chat` | 8x7B | 16384 | Chat Completion |
| `sonar-medium-online` | 8x7B | 12000 | Chat Completion |
| `codellama-34b-instruct`[3] | 34B | 16384 | Chat Completion |
| `codellama-70b-instruct` | 70B | 16384 | Chat Completion |
| `llama-2-70b-chat`[3] | 70B | 4096 | Chat Completion |
| `mistral-7b-instruct` [1] | 7B | 16384 | Chat Completion |
| `mixtral-8x7b-instruct` | 8x7B | 16384 | Chat Completion |
| `pplx-7b-chat`[2] [3] | 7B | 16384 | Chat Completion |
| `pplx-7b-online`[2] [3] | 7B | 12000 | Chat Completion |
| `pplx-70b-chat`[3] | 70B | 8192 | Chat Completion |
| `pplx-70b-online`[3] | 70B | 4000 | Chat Completion |
| Model | Parameter Count | Context Length | Model Type |
| :-------------------- | :-------------- | :------------- | :-------------- |
| `sonar-small-chat` | 7B | 16384 | Chat Completion |
| `sonar-small-online` | 7B | 12000 | Chat Completion |
| `sonar-medium-chat` | 8x7B | 16384 | Chat Completion |
| `sonar-medium-online` | 8x7B | 12000 | Chat Completion |
| `codellama-70b-instruct` | 70B | 16384 | Chat Completion |
| `mistral-7b-instruct` [1] | 7B | 16384 | Chat Completion |
| `mixtral-8x7b-instruct` | 8x7B | 16384 | Chat Completion |

View file

@ -8,7 +8,7 @@
// copy outputs into the export in ../models.js
// Update the date below if you run this again because Perplexity added new models.
// Last Collected: Feb 23, 2024
// Last Collected: Apr 14, 2024
import fs from "fs";

View file

@ -1,8 +1,26 @@
const MODELS = {
"togethercomputer/alpaca-7b": {
id: "togethercomputer/alpaca-7b",
organization: "Stanford",
name: "Alpaca (7B)",
"zero-one-ai/Yi-34B-Chat": {
id: "zero-one-ai/Yi-34B-Chat",
organization: "01.AI",
name: "01-ai Yi Chat (34B)",
maxLength: 4096,
},
"allenai/OLMo-7B-Instruct": {
id: "allenai/OLMo-7B-Instruct",
organization: "Allen AI",
name: "OLMo Instruct (7B)",
maxLength: 2048,
},
"allenai/OLMo-7B-Twin-2T": {
id: "allenai/OLMo-7B-Twin-2T",
organization: "Allen AI",
name: "OLMo Twin-2T (7B)",
maxLength: 2048,
},
"allenai/OLMo-7B": {
id: "allenai/OLMo-7B",
organization: "Allen AI",
name: "OLMo (7B)",
maxLength: 2048,
},
"Austism/chronos-hermes-13b": {
@ -11,96 +29,150 @@ const MODELS = {
name: "Chronos Hermes (13B)",
maxLength: 2048,
},
"togethercomputer/CodeLlama-13b-Instruct": {
id: "togethercomputer/CodeLlama-13b-Instruct",
organization: "Meta",
name: "Code Llama Instruct (13B)",
"cognitivecomputations/dolphin-2.5-mixtral-8x7b": {
id: "cognitivecomputations/dolphin-2.5-mixtral-8x7b",
organization: "cognitivecomputations",
name: "Dolphin 2.5 Mixtral 8x7b",
maxLength: 32768,
},
"databricks/dbrx-instruct": {
id: "databricks/dbrx-instruct",
organization: "databricks",
name: "DBRX Instruct",
maxLength: 32000,
},
"deepseek-ai/deepseek-coder-33b-instruct": {
id: "deepseek-ai/deepseek-coder-33b-instruct",
organization: "DeepSeek",
name: "Deepseek Coder Instruct (33B)",
maxLength: 16384,
},
"deepseek-ai/deepseek-llm-67b-chat": {
id: "deepseek-ai/deepseek-llm-67b-chat",
organization: "DeepSeek",
name: "DeepSeek LLM Chat (67B)",
maxLength: 4096,
},
"garage-bAInd/Platypus2-70B-instruct": {
id: "garage-bAInd/Platypus2-70B-instruct",
organization: "garage-bAInd",
name: "Platypus2 Instruct (70B)",
maxLength: 4096,
},
"google/gemma-2b-it": {
id: "google/gemma-2b-it",
organization: "Google",
name: "Gemma Instruct (2B)",
maxLength: 8192,
},
"togethercomputer/CodeLlama-34b-Instruct": {
id: "togethercomputer/CodeLlama-34b-Instruct",
organization: "Meta",
name: "Code Llama Instruct (34B)",
"google/gemma-7b-it": {
id: "google/gemma-7b-it",
organization: "Google",
name: "Gemma Instruct (7B)",
maxLength: 8192,
},
"togethercomputer/CodeLlama-7b-Instruct": {
id: "togethercomputer/CodeLlama-7b-Instruct",
organization: "Meta",
name: "Code Llama Instruct (7B)",
maxLength: 8192,
},
"DiscoResearch/DiscoLM-mixtral-8x7b-v2": {
id: "DiscoResearch/DiscoLM-mixtral-8x7b-v2",
organization: "DiscoResearch",
name: "DiscoLM Mixtral 8x7b",
maxLength: 32768,
},
"togethercomputer/falcon-40b-instruct": {
id: "togethercomputer/falcon-40b-instruct",
organization: "TII UAE",
name: "Falcon Instruct (40B)",
maxLength: 2048,
},
"togethercomputer/falcon-7b-instruct": {
id: "togethercomputer/falcon-7b-instruct",
organization: "TII UAE",
name: "Falcon Instruct (7B)",
maxLength: 2048,
},
"togethercomputer/GPT-NeoXT-Chat-Base-20B": {
id: "togethercomputer/GPT-NeoXT-Chat-Base-20B",
organization: "Together",
name: "GPT-NeoXT-Chat-Base (20B)",
maxLength: 2048,
},
"togethercomputer/llama-2-13b-chat": {
id: "togethercomputer/llama-2-13b-chat",
organization: "Meta",
name: "LLaMA-2 Chat (13B)",
maxLength: 4096,
},
"togethercomputer/llama-2-70b-chat": {
id: "togethercomputer/llama-2-70b-chat",
organization: "Meta",
name: "LLaMA-2 Chat (70B)",
maxLength: 4096,
},
"togethercomputer/llama-2-7b-chat": {
id: "togethercomputer/llama-2-7b-chat",
organization: "Meta",
name: "LLaMA-2 Chat (7B)",
maxLength: 4096,
},
"togethercomputer/Llama-2-7B-32K-Instruct": {
id: "togethercomputer/Llama-2-7B-32K-Instruct",
organization: "Together",
name: "LLaMA-2-7B-32K-Instruct (7B)",
maxLength: 32768,
},
"mistralai/Mistral-7B-Instruct-v0.1": {
id: "mistralai/Mistral-7B-Instruct-v0.1",
organization: "MistralAI",
name: "Mistral (7B) Instruct v0.1",
maxLength: 4096,
},
"mistralai/Mistral-7B-Instruct-v0.2": {
id: "mistralai/Mistral-7B-Instruct-v0.2",
organization: "MistralAI",
name: "Mistral (7B) Instruct v0.2",
maxLength: 32768,
},
"mistralai/Mixtral-8x7B-Instruct-v0.1": {
id: "mistralai/Mixtral-8x7B-Instruct-v0.1",
organization: "MistralAI",
name: "Mixtral-8x7B Instruct",
maxLength: 32768,
},
"Gryphe/MythoMax-L2-13b": {
id: "Gryphe/MythoMax-L2-13b",
organization: "Gryphe",
name: "MythoMax-L2 (13B)",
maxLength: 4096,
},
"lmsys/vicuna-13b-v1.5": {
id: "lmsys/vicuna-13b-v1.5",
organization: "LM Sys",
name: "Vicuna v1.5 (13B)",
maxLength: 4096,
},
"lmsys/vicuna-7b-v1.5": {
id: "lmsys/vicuna-7b-v1.5",
organization: "LM Sys",
name: "Vicuna v1.5 (7B)",
maxLength: 4096,
},
"codellama/CodeLlama-13b-Instruct-hf": {
id: "codellama/CodeLlama-13b-Instruct-hf",
organization: "Meta",
name: "Code Llama Instruct (13B)",
maxLength: 16384,
},
"codellama/CodeLlama-34b-Instruct-hf": {
id: "codellama/CodeLlama-34b-Instruct-hf",
organization: "Meta",
name: "Code Llama Instruct (34B)",
maxLength: 16384,
},
"codellama/CodeLlama-70b-Instruct-hf": {
id: "codellama/CodeLlama-70b-Instruct-hf",
organization: "Meta",
name: "Code Llama Instruct (70B)",
maxLength: 4096,
},
"codellama/CodeLlama-7b-Instruct-hf": {
id: "codellama/CodeLlama-7b-Instruct-hf",
organization: "Meta",
name: "Code Llama Instruct (7B)",
maxLength: 16384,
},
"meta-llama/Llama-2-70b-chat-hf": {
id: "meta-llama/Llama-2-70b-chat-hf",
organization: "Meta",
name: "LLaMA-2 Chat (70B)",
maxLength: 4096,
},
"meta-llama/Llama-2-13b-chat-hf": {
id: "meta-llama/Llama-2-13b-chat-hf",
organization: "Meta",
name: "LLaMA-2 Chat (13B)",
maxLength: 4096,
},
"meta-llama/Llama-2-7b-chat-hf": {
id: "meta-llama/Llama-2-7b-chat-hf",
organization: "Meta",
name: "LLaMA-2 Chat (7B)",
maxLength: 4096,
},
"mistralai/Mistral-7B-Instruct-v0.1": {
id: "mistralai/Mistral-7B-Instruct-v0.1",
organization: "mistralai",
name: "Mistral (7B) Instruct",
maxLength: 8192,
},
"mistralai/Mistral-7B-Instruct-v0.2": {
id: "mistralai/Mistral-7B-Instruct-v0.2",
organization: "mistralai",
name: "Mistral (7B) Instruct v0.2",
maxLength: 32768,
},
"mistralai/Mixtral-8x7B-Instruct-v0.1": {
id: "mistralai/Mixtral-8x7B-Instruct-v0.1",
organization: "mistralai",
name: "Mixtral-8x7B Instruct (46.7B)",
maxLength: 32768,
},
"NousResearch/Nous-Capybara-7B-V1p9": {
id: "NousResearch/Nous-Capybara-7B-V1p9",
organization: "NousResearch",
name: "Nous Capybara v1.9 (7B)",
maxLength: 8192,
},
"NousResearch/Nous-Hermes-2-Mistral-7B-DPO": {
id: "NousResearch/Nous-Hermes-2-Mistral-7B-DPO",
organization: "NousResearch",
name: "Nous Hermes 2 - Mistral DPO (7B)",
maxLength: 32768,
},
"NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO": {
id: "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO",
organization: "NousResearch",
name: "Nous Hermes 2 - Mixtral 8x7B-DPO (46.7B)",
maxLength: 32768,
},
"NousResearch/Nous-Hermes-2-Mixtral-8x7B-SFT": {
id: "NousResearch/Nous-Hermes-2-Mixtral-8x7B-SFT",
organization: "NousResearch",
name: "Nous Hermes 2 - Mixtral 8x7B-SFT (46.7B)",
maxLength: 32768,
},
"NousResearch/Nous-Hermes-llama-2-7b": {
id: "NousResearch/Nous-Hermes-llama-2-7b",
organization: "NousResearch",
@ -113,66 +185,96 @@ const MODELS = {
name: "Nous Hermes Llama-2 (13B)",
maxLength: 4096,
},
"NousResearch/Nous-Hermes-Llama2-70b": {
id: "NousResearch/Nous-Hermes-Llama2-70b",
organization: "NousResearch",
name: "Nous Hermes Llama-2 (70B)",
maxLength: 4096,
},
"NousResearch/Nous-Hermes-2-Yi-34B": {
id: "NousResearch/Nous-Hermes-2-Yi-34B",
organization: "NousResearch",
name: "Nous Hermes-2 Yi (34B)",
maxLength: 4096,
},
"NousResearch/Nous-Capybara-7B-V1p9": {
id: "NousResearch/Nous-Capybara-7B-V1p9",
organization: "NousResearch",
name: "Nous Capybara v1.9 (7B)",
maxLength: 8192,
},
"openchat/openchat-3.5-1210": {
id: "openchat/openchat-3.5-1210",
organization: "OpenChat",
name: "OpenChat 3.5 1210 (7B)",
name: "OpenChat 3.5 (7B)",
maxLength: 8192,
},
"teknium/OpenHermes-2-Mistral-7B": {
id: "teknium/OpenHermes-2-Mistral-7B",
organization: "teknium",
name: "OpenHermes-2-Mistral (7B)",
maxLength: 4096,
},
"teknium/OpenHermes-2p5-Mistral-7B": {
id: "teknium/OpenHermes-2p5-Mistral-7B",
organization: "teknium",
name: "OpenHermes-2.5-Mistral (7B)",
maxLength: 4096,
},
"Open-Orca/Mistral-7B-OpenOrca": {
id: "Open-Orca/Mistral-7B-OpenOrca",
organization: "OpenOrca",
name: "OpenOrca Mistral (7B) 8K",
maxLength: 8192,
},
"garage-bAInd/Platypus2-70B-instruct": {
id: "garage-bAInd/Platypus2-70B-instruct",
organization: "garage-bAInd",
name: "Platypus2 Instruct (70B)",
maxLength: 4096,
"Qwen/Qwen1.5-0.5B-Chat": {
id: "Qwen/Qwen1.5-0.5B-Chat",
organization: "Qwen",
name: "Qwen 1.5 Chat (0.5B)",
maxLength: 32768,
},
"togethercomputer/Pythia-Chat-Base-7B-v0.16": {
id: "togethercomputer/Pythia-Chat-Base-7B-v0.16",
organization: "Together",
name: "Pythia-Chat-Base (7B)",
"Qwen/Qwen1.5-1.8B-Chat": {
id: "Qwen/Qwen1.5-1.8B-Chat",
organization: "Qwen",
name: "Qwen 1.5 Chat (1.8B)",
maxLength: 32768,
},
"Qwen/Qwen1.5-4B-Chat": {
id: "Qwen/Qwen1.5-4B-Chat",
organization: "Qwen",
name: "Qwen 1.5 Chat (4B)",
maxLength: 32768,
},
"Qwen/Qwen1.5-7B-Chat": {
id: "Qwen/Qwen1.5-7B-Chat",
organization: "Qwen",
name: "Qwen 1.5 Chat (7B)",
maxLength: 32768,
},
"Qwen/Qwen1.5-14B-Chat": {
id: "Qwen/Qwen1.5-14B-Chat",
organization: "Qwen",
name: "Qwen 1.5 Chat (14B)",
maxLength: 32768,
},
"Qwen/Qwen1.5-32B-Chat": {
id: "Qwen/Qwen1.5-32B-Chat",
organization: "Qwen",
name: "Qwen 1.5 Chat (32B)",
maxLength: 32768,
},
"Qwen/Qwen1.5-72B-Chat": {
id: "Qwen/Qwen1.5-72B-Chat",
organization: "Qwen",
name: "Qwen 1.5 Chat (72B)",
maxLength: 32768,
},
"snorkelai/Snorkel-Mistral-PairRM-DPO": {
id: "snorkelai/Snorkel-Mistral-PairRM-DPO",
organization: "Snorkel AI",
name: "Snorkel Mistral PairRM DPO (7B)",
maxLength: 32768,
},
"togethercomputer/alpaca-7b": {
id: "togethercomputer/alpaca-7b",
organization: "Stanford",
name: "Alpaca (7B)",
maxLength: 2048,
},
"togethercomputer/Qwen-7B-Chat": {
id: "togethercomputer/Qwen-7B-Chat",
organization: "Qwen",
name: "Qwen-Chat (7B)",
"teknium/OpenHermes-2-Mistral-7B": {
id: "teknium/OpenHermes-2-Mistral-7B",
organization: "Teknium",
name: "OpenHermes-2-Mistral (7B)",
maxLength: 8192,
},
"teknium/OpenHermes-2p5-Mistral-7B": {
id: "teknium/OpenHermes-2p5-Mistral-7B",
organization: "Teknium",
name: "OpenHermes-2.5-Mistral (7B)",
maxLength: 8192,
},
"togethercomputer/Llama-2-7B-32K-Instruct": {
id: "togethercomputer/Llama-2-7B-32K-Instruct",
organization: "Together",
name: "LLaMA-2-7B-32K-Instruct (7B)",
maxLength: 32768,
},
"togethercomputer/RedPajama-INCITE-Chat-3B-v1": {
id: "togethercomputer/RedPajama-INCITE-Chat-3B-v1",
organization: "Together",
@ -185,40 +287,34 @@ const MODELS = {
name: "RedPajama-INCITE Chat (7B)",
maxLength: 2048,
},
"upstage/SOLAR-0-70b-16bit": {
id: "upstage/SOLAR-0-70b-16bit",
organization: "Upstage",
name: "SOLAR v0 (70B)",
maxLength: 4096,
},
"togethercomputer/StripedHyena-Nous-7B": {
id: "togethercomputer/StripedHyena-Nous-7B",
organization: "Together",
name: "StripedHyena Nous (7B)",
maxLength: 32768,
},
"lmsys/vicuna-7b-v1.5": {
id: "lmsys/vicuna-7b-v1.5",
organization: "LM Sys",
name: "Vicuna v1.5 (7B)",
"Undi95/ReMM-SLERP-L2-13B": {
id: "Undi95/ReMM-SLERP-L2-13B",
organization: "Undi95",
name: "ReMM SLERP L2 (13B)",
maxLength: 4096,
},
"lmsys/vicuna-13b-v1.5": {
id: "lmsys/vicuna-13b-v1.5",
organization: "LM Sys",
name: "Vicuna v1.5 (13B)",
"Undi95/Toppy-M-7B": {
id: "Undi95/Toppy-M-7B",
organization: "Undi95",
name: "Toppy M (7B)",
maxLength: 4096,
},
"lmsys/vicuna-13b-v1.5-16k": {
id: "lmsys/vicuna-13b-v1.5-16k",
organization: "LM Sys",
name: "Vicuna v1.5 16K (13B)",
maxLength: 16384,
"WizardLM/WizardLM-13B-V1.2": {
id: "WizardLM/WizardLM-13B-V1.2",
organization: "WizardLM",
name: "WizardLM v1.2 (13B)",
maxLength: 4096,
},
"zero-one-ai/Yi-34B-Chat": {
id: "zero-one-ai/Yi-34B-Chat",
organization: "01.AI",
name: "01-ai Yi Chat (34B)",
"upstage/SOLAR-10.7B-Instruct-v1.0": {
id: "upstage/SOLAR-10.7B-Instruct-v1.0",
organization: "upstage",
name: "Upstage SOLAR Instruct v1 (11B)",
maxLength: 4096,
},
};

View file

@ -1,39 +1,55 @@
| Organization | Model Name | Model String for API | Max Seq Length |
| ------------- | ---------------------------- | -------------------------------------------- | -------------- |
| Stanford | Alpaca (7B) | togethercomputer/alpaca-7b | 2048 |
| Austism | Chronos Hermes (13B) | Austism/chronos-hermes-13b | 2048 |
| Meta | Code Llama Instruct (13B) | togethercomputer/CodeLlama-13b-Instruct | 8192 |
| Meta | Code Llama Instruct (34B) | togethercomputer/CodeLlama-34b-Instruct | 8192 |
| Meta | Code Llama Instruct (7B) | togethercomputer/CodeLlama-7b-Instruct | 8192 |
| DiscoResearch | DiscoLM Mixtral 8x7b | DiscoResearch/DiscoLM-mixtral-8x7b-v2 | 32768 |
| TII UAE | Falcon Instruct (40B) | togethercomputer/falcon-40b-instruct | 2048 |
| TII UAE | Falcon Instruct (7B) | togethercomputer/falcon-7b-instruct | 2048 |
| Together | GPT-NeoXT-Chat-Base (20B) | togethercomputer/GPT-NeoXT-Chat-Base-20B | 2048 |
| Meta | LLaMA-2 Chat (13B) | togethercomputer/llama-2-13b-chat | 4096 |
| Meta | LLaMA-2 Chat (70B) | togethercomputer/llama-2-70b-chat | 4096 |
| Meta | LLaMA-2 Chat (7B) | togethercomputer/llama-2-7b-chat | 4096 |
| Together | LLaMA-2-7B-32K-Instruct (7B) | togethercomputer/Llama-2-7B-32K-Instruct | 32768 |
| MistralAI | Mistral (7B) Instruct v0.1 | mistralai/Mistral-7B-Instruct-v0.1 | 4096 |
| MistralAI | Mistral (7B) Instruct v0.2 | mistralai/Mistral-7B-Instruct-v0.2 | 32768 |
| MistralAI | Mixtral-8x7B Instruct | mistralai/Mixtral-8x7B-Instruct-v0.1 | 32768 |
| Gryphe | MythoMax-L2 (13B) | Gryphe/MythoMax-L2-13b | 4096 |
| NousResearch | Nous Hermes LLaMA-2 (7B) | NousResearch/Nous-Hermes-llama-2-7b | 4096 |
| NousResearch | Nous Hermes Llama-2 (13B) | NousResearch/Nous-Hermes-Llama2-13b | 4096 |
| NousResearch | Nous Hermes Llama-2 (70B) | NousResearch/Nous-Hermes-Llama2-70b | 4096 |
| NousResearch | Nous Hermes-2 Yi (34B) | NousResearch/Nous-Hermes-2-Yi-34B | 4096 |
| NousResearch | Nous Capybara v1.9 (7B) | NousResearch/Nous-Capybara-7B-V1p9 | 8192 |
| OpenChat | OpenChat 3.5 1210 (7B) | openchat/openchat-3.5-1210 | 8192 |
| teknium | OpenHermes-2-Mistral (7B) | teknium/OpenHermes-2-Mistral-7B | 4096 |
| teknium | OpenHermes-2.5-Mistral (7B) | teknium/OpenHermes-2p5-Mistral-7B | 4096 |
| OpenOrca | OpenOrca Mistral (7B) 8K | Open-Orca/Mistral-7B-OpenOrca | 8192 |
| garage-bAInd | Platypus2 Instruct (70B) | garage-bAInd/Platypus2-70B-instruct | 4096 |
| Together | Pythia-Chat-Base (7B) | togethercomputer/Pythia-Chat-Base-7B-v0.16 | 2048 |
| Qwen | Qwen-Chat (7B) | togethercomputer/Qwen-7B-Chat | 8192 |
| Together | RedPajama-INCITE Chat (3B) | togethercomputer/RedPajama-INCITE-Chat-3B-v1 | 2048 |
| Together | RedPajama-INCITE Chat (7B) | togethercomputer/RedPajama-INCITE-7B-Chat | 2048 |
| Upstage | SOLAR v0 (70B) | upstage/SOLAR-0-70b-16bit | 4096 |
| Together | StripedHyena Nous (7B) | togethercomputer/StripedHyena-Nous-7B | 32768 |
| LM Sys | Vicuna v1.5 (7B) | lmsys/vicuna-7b-v1.5 | 4096 |
| LM Sys | Vicuna v1.5 (13B) | lmsys/vicuna-13b-v1.5 | 4096 |
| LM Sys | Vicuna v1.5 16K (13B) | lmsys/vicuna-13b-v1.5-16k | 16384 |
| 01.AI | 01-ai Yi Chat (34B) | zero-one-ai/Yi-34B-Chat | 4096 |
| Organization | Model Name | Model String for API | Context length |
| --- | --- | --- | --- |
| 01.AI | 01-ai Yi Chat (34B) | zero-one-ai/Yi-34B-Chat | 4096 |
| Allen AI | OLMo Instruct (7B) | allenai/OLMo-7B-Instruct | 2048 |
| Allen AI | OLMo Twin-2T (7B) | allenai/OLMo-7B-Twin-2T | 2048 |
| Allen AI | OLMo (7B) | allenai/OLMo-7B | 2048 |
| Austism | Chronos Hermes (13B) | Austism/chronos-hermes-13b | 2048 |
| cognitivecomputations | Dolphin 2.5 Mixtral 8x7b | cognitivecomputations/dolphin-2.5-mixtral-8x7b | 32768 |
| databricks | DBRX Instruct | databricks/dbrx-instruct | 32000 |
| DeepSeek | Deepseek Coder Instruct (33B) | deepseek-ai/deepseek-coder-33b-instruct | 16384 |
| DeepSeek | DeepSeek LLM Chat (67B) | deepseek-ai/deepseek-llm-67b-chat | 4096 |
| garage-bAInd | Platypus2 Instruct (70B) | garage-bAInd/Platypus2-70B-instruct | 4096 |
| Google | Gemma Instruct (2B) | google/gemma-2b-it | 8192 |
| Google | Gemma Instruct (7B) | google/gemma-7b-it | 8192 |
| Gryphe | MythoMax-L2 (13B) | Gryphe/MythoMax-L2-13b | 4096 |
| LM Sys | Vicuna v1.5 (13B) | lmsys/vicuna-13b-v1.5 | 4096 |
| LM Sys | Vicuna v1.5 (7B) | lmsys/vicuna-7b-v1.5 | 4096 |
| Meta | Code Llama Instruct (13B) | codellama/CodeLlama-13b-Instruct-hf | 16384 |
| Meta | Code Llama Instruct (34B) | codellama/CodeLlama-34b-Instruct-hf | 16384 |
| Meta | Code Llama Instruct (70B) | codellama/CodeLlama-70b-Instruct-hf | 4096 |
| Meta | Code Llama Instruct (7B) | codellama/CodeLlama-7b-Instruct-hf | 16384 |
| Meta | LLaMA-2 Chat (70B) | meta-llama/Llama-2-70b-chat-hf | 4096 |
| Meta | LLaMA-2 Chat (13B) | meta-llama/Llama-2-13b-chat-hf | 4096 |
| Meta | LLaMA-2 Chat (7B) | meta-llama/Llama-2-7b-chat-hf | 4096 |
| mistralai | Mistral (7B) Instruct | mistralai/Mistral-7B-Instruct-v0.1 | 8192 |
| mistralai | Mistral (7B) Instruct v0.2 | mistralai/Mistral-7B-Instruct-v0.2 | 32768 |
| mistralai | Mixtral-8x7B Instruct (46.7B) | mistralai/Mixtral-8x7B-Instruct-v0.1 | 32768 |
| NousResearch | Nous Capybara v1.9 (7B) | NousResearch/Nous-Capybara-7B-V1p9 | 8192 |
| NousResearch | Nous Hermes 2 - Mistral DPO (7B) | NousResearch/Nous-Hermes-2-Mistral-7B-DPO | 32768 |
| NousResearch | Nous Hermes 2 - Mixtral 8x7B-DPO (46.7B) | NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO | 32768 |
| NousResearch | Nous Hermes 2 - Mixtral 8x7B-SFT (46.7B) | NousResearch/Nous-Hermes-2-Mixtral-8x7B-SFT | 32768 |
| NousResearch | Nous Hermes LLaMA-2 (7B) | NousResearch/Nous-Hermes-llama-2-7b | 4096 |
| NousResearch | Nous Hermes Llama-2 (13B) | NousResearch/Nous-Hermes-Llama2-13b | 4096 |
| NousResearch | Nous Hermes-2 Yi (34B) | NousResearch/Nous-Hermes-2-Yi-34B | 4096 |
| OpenChat | OpenChat 3.5 (7B) | openchat/openchat-3.5-1210 | 8192 |
| OpenOrca | OpenOrca Mistral (7B) 8K | Open-Orca/Mistral-7B-OpenOrca | 8192 |
| Qwen | Qwen 1.5 Chat (0.5B) | Qwen/Qwen1.5-0.5B-Chat | 32768 |
| Qwen | Qwen 1.5 Chat (1.8B) | Qwen/Qwen1.5-1.8B-Chat | 32768 |
| Qwen | Qwen 1.5 Chat (4B) | Qwen/Qwen1.5-4B-Chat | 32768 |
| Qwen | Qwen 1.5 Chat (7B) | Qwen/Qwen1.5-7B-Chat | 32768 |
| Qwen | Qwen 1.5 Chat (14B) | Qwen/Qwen1.5-14B-Chat | 32768 |
| Qwen | Qwen 1.5 Chat (32B) | Qwen/Qwen1.5-32B-Chat | 32768 |
| Qwen | Qwen 1.5 Chat (72B) | Qwen/Qwen1.5-72B-Chat | 32768 |
| Snorkel AI | Snorkel Mistral PairRM DPO (7B) | snorkelai/Snorkel-Mistral-PairRM-DPO | 32768 |
| Stanford | Alpaca (7B) | togethercomputer/alpaca-7b | 2048 |
| Teknium | OpenHermes-2-Mistral (7B) | teknium/OpenHermes-2-Mistral-7B | 8192 |
| Teknium | OpenHermes-2.5-Mistral (7B) | teknium/OpenHermes-2p5-Mistral-7B | 8192 |
| Together | LLaMA-2-7B-32K-Instruct (7B) | togethercomputer/Llama-2-7B-32K-Instruct | 32768 |
| Together | RedPajama-INCITE Chat (3B) | togethercomputer/RedPajama-INCITE-Chat-3B-v1 | 2048 |
| Together | RedPajama-INCITE Chat (7B) | togethercomputer/RedPajama-INCITE-7B-Chat | 2048 |
| Together | StripedHyena Nous (7B) | togethercomputer/StripedHyena-Nous-7B | 32768 |
| Undi95 | ReMM SLERP L2 (13B) | Undi95/ReMM-SLERP-L2-13B | 4096 |
| Undi95 | Toppy M (7B) | Undi95/Toppy-M-7B | 4096 |
| WizardLM | WizardLM v1.2 (13B) | WizardLM/WizardLM-13B-V1.2 | 4096 |
| upstage | Upstage SOLAR Instruct v1 (11B) | upstage/SOLAR-10.7B-Instruct-v1.0 | 4096 |

View file

@ -8,7 +8,9 @@
// copy outputs into the export in ../models.js
// Update the date below if you run this again because TogetherAI added new models.
// Last Collected: Jan 10, 2023
// Last Collected: Apr 14, 2024
// Since last collection Together's docs are broken. I just copied the HTML table
// and had claude3 convert to markdown and it works well enough.
import fs from "fs";