From a385ea3d82ddf0c822455169b6ef840c42662c41 Mon Sep 17 00:00:00 2001 From: Timothy Carambat <rambat1010@gmail.com> Date: Fri, 23 Feb 2024 17:33:16 -0800 Subject: [PATCH] CHORE: bump pplx model support (#791) bump pplx model support --- server/utils/AiProviders/perplexity/models.js | 38 ++++++++++++++----- .../perplexity/scripts/chat_models.txt | 26 +++++++------ .../AiProviders/perplexity/scripts/parse.mjs | 4 +- 3 files changed, 46 insertions(+), 22 deletions(-) diff --git a/server/utils/AiProviders/perplexity/models.js b/server/utils/AiProviders/perplexity/models.js index 258cfeace..95cd8eac7 100644 --- a/server/utils/AiProviders/perplexity/models.js +++ b/server/utils/AiProviders/perplexity/models.js @@ -1,4 +1,24 @@ const MODELS = { + "sonar-small-chat": { + id: "sonar-small-chat", + name: "sonar-small-chat", + maxLength: 16384, + }, + "sonar-small-online": { + id: "sonar-small-online", + name: "sonar-small-online", + maxLength: 12000, + }, + "sonar-medium-chat": { + id: "sonar-medium-chat", + name: "sonar-medium-chat", + maxLength: 16384, + }, + "sonar-medium-online": { + id: "sonar-medium-online", + name: "sonar-medium-online", + maxLength: 12000, + }, "codellama-34b-instruct": { id: "codellama-34b-instruct", name: "codellama-34b-instruct", @@ -17,32 +37,32 @@ const MODELS = { "mistral-7b-instruct": { id: "mistral-7b-instruct", name: "mistral-7b-instruct", - maxLength: 8192, + maxLength: 16384, }, "mixtral-8x7b-instruct": { id: "mixtral-8x7b-instruct", name: "mixtral-8x7b-instruct", - maxLength: 8192, + maxLength: 16384, }, "pplx-7b-chat": { id: "pplx-7b-chat", name: "pplx-7b-chat", - maxLength: 8192, + maxLength: 16384, + }, + "pplx-7b-online": { + id: "pplx-7b-online", + name: "pplx-7b-online", + maxLength: 12000, }, "pplx-70b-chat": { id: "pplx-70b-chat", name: "pplx-70b-chat", maxLength: 8192, }, - "pplx-7b-online": { - id: "pplx-7b-online", - name: "pplx-7b-online", - maxLength: 8192, - }, "pplx-70b-online": { id: "pplx-70b-online", name: "pplx-70b-online", - maxLength: 8192, + maxLength: 4000, }, }; diff --git a/server/utils/AiProviders/perplexity/scripts/chat_models.txt b/server/utils/AiProviders/perplexity/scripts/chat_models.txt index 83f6d2a80..97ba9017a 100644 --- a/server/utils/AiProviders/perplexity/scripts/chat_models.txt +++ b/server/utils/AiProviders/perplexity/scripts/chat_models.txt @@ -1,11 +1,15 @@ -| Model | Context Length | Model Type | -| :------------------------ | :------------- | :-------------- | -| `codellama-34b-instruct` | 16384 | Chat Completion | -| `codellama-70b-instruct` | 16384 | Chat Completion | -| `llama-2-70b-chat` | 4096 | Chat Completion | -| `mistral-7b-instruct` [2] | 8192 [1] | Chat Completion | -| `mixtral-8x7b-instruct` | 8192 [1] | Chat Completion | -| `pplx-7b-chat` | 8192 | Chat Completion | -| `pplx-70b-chat` | 8192 | Chat Completion | -| `pplx-7b-online` | 8192 | Chat Completion | -| `pplx-70b-online` | 8192 | Chat Completion | \ No newline at end of file +| Model | Parameter Count | Context Length | Model Type | +| :-------------------------- | :-------------- | :------------- | :-------------- | +| `sonar-small-chat` | 7B | 16384 | Chat Completion | +| `sonar-small-online` | 7B | 12000 | Chat Completion | +| `sonar-medium-chat` | 8x7B | 16384 | Chat Completion | +| `sonar-medium-online` | 8x7B | 12000 | Chat Completion | +| `codellama-34b-instruct`[3] | 34B | 16384 | Chat Completion | +| `codellama-70b-instruct` | 70B | 16384 | Chat Completion | +| `llama-2-70b-chat`[3] | 70B | 4096 | Chat Completion | +| `mistral-7b-instruct` [1] | 7B | 16384 | Chat Completion | +| `mixtral-8x7b-instruct` | 8x7B | 16384 | Chat Completion | +| `pplx-7b-chat`[2] [3] | 7B | 16384 | Chat Completion | +| `pplx-7b-online`[2] [3] | 7B | 12000 | Chat Completion | +| `pplx-70b-chat`[3] | 70B | 8192 | Chat Completion | +| `pplx-70b-online`[3] | 70B | 4000 | Chat Completion | \ No newline at end of file diff --git a/server/utils/AiProviders/perplexity/scripts/parse.mjs b/server/utils/AiProviders/perplexity/scripts/parse.mjs index 749a63dce..d2064354a 100644 --- a/server/utils/AiProviders/perplexity/scripts/parse.mjs +++ b/server/utils/AiProviders/perplexity/scripts/parse.mjs @@ -8,7 +8,7 @@ // copy outputs into the export in ../models.js // Update the date below if you run this again because Perplexity added new models. -// Last Collected: Feb 22, 2024 +// Last Collected: Feb 23, 2024 import fs from "fs"; @@ -18,7 +18,7 @@ function parseChatModels() { const rows = tableString.split("\n").slice(2); rows.forEach((row) => { - let [model, contextLength] = row + let [model, _, contextLength] = row .split("|") .slice(1, -1) .map((text) => text.trim());