Timothy Carambat
dd7c4675d3
LLM performance metric tracking ( #2825 )
...
* WIP performance metric tracking
* fix: patch UI trying to .toFixed() null metric
Anthropic tracking migraiton
cleanup logs
* Apipie implmentation, not tested
* Cleanup Anthropic notes, Add support for AzureOpenAI tracking
* bedrock token metric tracking
* Cohere support
* feat: improve default stream handler to track for provider who are actually OpenAI compliant in usage reporting
add deepseek support
* feat: Add FireworksAI tracking reporting
fix: improve handler when usage:null is reported (why?)
* Add token reporting for GenericOpenAI
* token reporting for koboldcpp + lmstudio
* lint
* support Groq token tracking
* HF token tracking
* token tracking for togetherai
* LiteLLM token tracking
* linting + Mitral token tracking support
* XAI token metric reporting
* native provider runner
* LocalAI token tracking
* Novita token tracking
* OpenRouter token tracking
* Apipie stream metrics
* textwebgenui token tracking
* perplexity token reporting
* ollama token reporting
* lint
* put back comment
* Rip out LC ollama wrapper and use official library
* patch images with new ollama lib
* improve ollama offline message
* fix image handling in ollama llm provider
* lint
* NVIDIA NIM token tracking
* update openai compatbility responses
* UI/UX show/hide metrics on click for user preference
* update bedrock client
---------
Co-authored-by: shatfield4 <seanhatfield5@gmail.com>
2024-12-16 14:31:17 -08:00
Timothy Carambat
906eb70ca1
bump Perplexity models ( #2275 )
2024-09-12 13:13:47 -07:00
timothycarambat
cb7cb2d976
Add 405B to perplexity
2024-08-19 12:26:22 -07:00
Timothy Carambat
99f2c25b1c
Agent Context window + context window refactor. ( #2126 )
...
* Enable agent context windows to be accurate per provider:model
* Refactor model mapping to external file
Add token count to document length instead of char-count
refernce promptWindowLimit from AIProvider in central location
* remove unused imports
2024-08-15 12:13:28 -07:00
timothycarambat
466bf7dc9c
Bump Perplexity and Together AI static model list
2024-07-31 10:58:34 -07:00
timothycarambat
296f041564
path perplexity model ids
...
closes #1990
2024-07-28 16:29:18 -07:00
Timothy Carambat
5df6b5f7d9
Bump perplexity models ( #1905 )
...
* Added Supported Models Free Tier - chat_models.txt
Need to fill in correct Parameter Count.
* Bump perplexity model
closes #1901
closes #1900
---------
Co-authored-by: Tim-Hoekstra <135951177+Tim-Hoekstra@users.noreply.github.com>
2024-07-19 15:11:10 -07:00
Timothy Carambat
0b845fbb1c
Deprecate .isSafe
moderation ( #1790 )
...
Add type defs to helpers
2024-06-28 15:32:30 -07:00
timothycarambat
9f327d015a
update error handling for OpenAI providers
2024-05-22 09:58:10 -05:00
Timothy Carambat
01cf2fed17
Make native embedder the fallback for all LLMs ( #1427 )
2024-05-16 17:25:05 -07:00
Sean Hatfield
9feaad79cc
[CHORE] Remove sendChat and streamChat in all LLM providers ( #1260 )
...
* remove sendChat and streamChat functions/references in all LLM providers
* remove unused imports
---------
Co-authored-by: timothycarambat <rambat1010@gmail.com>
2024-05-01 16:52:28 -07:00
Timothy Carambat
547d4859ef
Bump openai
package to latest ( #1234 )
...
* Bump `openai` package to latest
Tested all except localai
* bump LocalAI support with latest image
* add deprecation notice
* linting
2024-04-30 12:33:42 -07:00
timothycarambat
dfaaf1680f
update perplexity models
...
resolves #1188
2024-04-25 07:34:28 -07:00
Timothy Carambat
8306098b08
Bump all static model providers ( #1101 )
2024-04-14 12:55:21 -07:00
Timothy Carambat
a385ea3d82
CHORE: bump pplx model support ( #791 )
...
bump pplx model support
2024-02-23 17:33:16 -08:00
Sean Hatfield
80ced5eba4
[FEAT] PerplexityAI Support ( #778 )
...
* add LLM support for perplexity
* update README & example env
* fix ENV keys in example env files
* slight changes for QA of perplexity support
* Update Perplexity AI name
---------
Co-authored-by: timothycarambat <rambat1010@gmail.com>
2024-02-22 12:48:57 -08:00