Timothy Carambat
c4f75feb08
Support historical message image inputs/attachments for n+1 queries ( #2919 )
...
* Support historical message image inputs/attachments for n+1 queries
* patch gemini
* OpenRouter vision support cleanup
* xai vision history support
* Mistral logging
---------
Co-authored-by: shatfield4 <seanhatfield5@gmail.com>
2025-01-16 13:49:06 -08:00
Timothy Carambat
21af81085a
Add caching to Gemini /models ( #2969 )
...
rename file typo
2025-01-13 13:12:03 -08:00
Timothy Carambat
b082c8e441
Add support for gemini authenticated models endpoint ( #2868 )
...
* Add support for gemini authenticated models endpoint
add customModels entry
add un-authed fallback to default listing
separate models by expiermental status
resolves #2866
* add back improved logic for apiVersion decision making
2024-12-17 15:20:26 -08:00
Timothy Carambat
dd7c4675d3
LLM performance metric tracking ( #2825 )
...
* WIP performance metric tracking
* fix: patch UI trying to .toFixed() null metric
Anthropic tracking migraiton
cleanup logs
* Apipie implmentation, not tested
* Cleanup Anthropic notes, Add support for AzureOpenAI tracking
* bedrock token metric tracking
* Cohere support
* feat: improve default stream handler to track for provider who are actually OpenAI compliant in usage reporting
add deepseek support
* feat: Add FireworksAI tracking reporting
fix: improve handler when usage:null is reported (why?)
* Add token reporting for GenericOpenAI
* token reporting for koboldcpp + lmstudio
* lint
* support Groq token tracking
* HF token tracking
* token tracking for togetherai
* LiteLLM token tracking
* linting + Mitral token tracking support
* XAI token metric reporting
* native provider runner
* LocalAI token tracking
* Novita token tracking
* OpenRouter token tracking
* Apipie stream metrics
* textwebgenui token tracking
* perplexity token reporting
* ollama token reporting
* lint
* put back comment
* Rip out LC ollama wrapper and use official library
* patch images with new ollama lib
* improve ollama offline message
* fix image handling in ollama llm provider
* lint
* NVIDIA NIM token tracking
* update openai compatbility responses
* UI/UX show/hide metrics on click for user preference
* update bedrock client
---------
Co-authored-by: shatfield4 <seanhatfield5@gmail.com>
2024-12-16 14:31:17 -08:00
timothycarambat
f8e91e1ffa
patch gemini-2.0-key
2024-12-11 16:52:31 -08:00
timothycarambat
69b672b625
add gemini 1206 and gemini-2.0-flash exp models
...
connect #2788
2024-12-11 09:04:29 -08:00
timothycarambat
62be0cd0c5
add gemini-exp-1121 to expiermental set
2024-11-22 09:36:44 -08:00
timothycarambat
246152c024
Add gemini-exp-1121
...
resolves #2657
2024-11-21 11:02:43 -08:00
Timothy Carambat
26e2d8cc3b
Add more expiermental models from Gemini ( #2663 )
2024-11-20 09:52:33 -08:00
Timothy Carambat
c612239ecb
Add Gemini exp
models ( #2268 )
...
Add Gemini models
resolves #2263
2024-09-11 13:03:14 -07:00
Timothy Carambat
99f2c25b1c
Agent Context window + context window refactor. ( #2126 )
...
* Enable agent context windows to be accurate per provider:model
* Refactor model mapping to external file
Add token count to document length instead of char-count
refernce promptWindowLimit from AIProvider in central location
* remove unused imports
2024-08-15 12:13:28 -07:00
Timothy Carambat
f06ef6180d
add exp model to v1Beta ( #2082 )
2024-08-09 14:19:49 -07:00
RahSwe
c55ef33fce
Gemini Pro 1.5, API support for 2M context and new experimental model ( #2031 )
2024-08-02 10:24:31 -07:00
Timothy Carambat
38fc181238
Add multimodality support ( #2001 )
...
* Add multimodality support
* Add Bedrock, KoboldCpp,LocalAI,and TextWebGenUI multi-modal
* temp dev build
* patch bad import
* noscrolls for windows dnd
* noscrolls for windows dnd
* update README
* update README
* add multimodal check
2024-07-31 10:47:49 -07:00
Timothy Carambat
0b845fbb1c
Deprecate .isSafe
moderation ( #1790 )
...
Add type defs to helpers
2024-06-28 15:32:30 -07:00
Sean Hatfield
3f78ef413b
[FEAT] Support for gemini-1.0-pro model and fixes to prompt window limit ( #1557 )
...
support for gemini-1.0-pro model and fixes to prompt window limit
2024-05-29 08:17:35 +08:00
Sean Hatfield
cc7e7fb3ac
[FEAT] Add support for gemini-1.5-flash-latest model ( #1502 )
...
* add support for gemini-1.5-flash-latest
* update comment in gemini LLM provider
2024-05-23 09:42:30 -07:00
Timothy Carambat
28eba636e9
Allow setting of safety thresholds for Gemini ( #1466 )
...
* Allow setting of safety thresholds for Gemini
* linting
2024-05-20 13:17:00 -05:00
Timothy Carambat
01cf2fed17
Make native embedder the fallback for all LLMs ( #1427 )
2024-05-16 17:25:05 -07:00
Sean Hatfield
948ac8a3dd
[FIX] Validate messages schema for gemini provider ( #1351 )
...
validate messages schema for gemini provider
2024-05-10 17:33:25 -07:00
Sean Hatfield
9feaad79cc
[CHORE] Remove sendChat and streamChat in all LLM providers ( #1260 )
...
* remove sendChat and streamChat functions/references in all LLM providers
* remove unused imports
---------
Co-authored-by: timothycarambat <rambat1010@gmail.com>
2024-05-01 16:52:28 -07:00
Timothy Carambat
58b744771f
Add support for Gemini-1.5 Pro ( #1134 )
...
* Add support for Gemini-1.5 Pro
bump @google/generative-ai pkg
Toggle apiVersion if beta model selected
resolves #1109
* update API messages due to package change
2024-04-19 08:59:46 -07:00
Timothy Carambat
0dd6001fa6
Patch Gemini/Google AI errors ( #977 )
2024-03-26 17:20:12 -07:00
Timothy Carambat
0e46a11cb6
Stop generation button during stream-response ( #892 )
...
* Stop generation button during stream-response
* add custom stop icon
* add stop to thread chats
2024-03-12 15:21:27 -07:00
Timothy Carambat
c59ab9da0a
Refactor LLM chat backend ( #717 )
...
* refactor stream/chat/embed-stram to be a single execution logic path so that it is easier to maintain and build upon
* no thread in sync chat since only api uses it
adjust import locations
2024-02-14 12:32:07 -08:00
Timothy Carambat
aca5940650
Refactor handleStream to LLM Classes ( #685 )
2024-02-07 08:15:14 -08:00
Sean Hatfield
c2c8fe9756
add support for mistral api ( #610 )
...
* add support for mistral api
* update docs to show support for Mistral
* add default temp to all providers, suggest different results per provider
---------
Co-authored-by: timothycarambat <rambat1010@gmail.com>
2024-01-17 14:42:05 -08:00
Sean Hatfield
90df37582b
Per workspace model selection ( #582 )
...
* WIP model selection per workspace (migrations and openai saves properly
* revert OpenAiOption
* add support for models per workspace for anthropic, localAi, ollama, openAi, and togetherAi
* remove unneeded comments
* update logic for when LLMProvider is reset, reset Ai provider files with master
* remove frontend/api reset of workspace chat and move logic to updateENV
add postUpdate callbacks to envs
* set preferred model for chat on class instantiation
* remove extra param
* linting
* remove unused var
* refactor chat model selection on workspace
* linting
* add fallback for base path to localai models
---------
Co-authored-by: timothycarambat <rambat1010@gmail.com>
2024-01-17 12:59:25 -08:00
Timothy Carambat
6d5968bf7e
Llm chore cleanup ( #501 )
...
* move internal functions to private in class
simplify lc message convertor
* Fix hanging Context text when none is present
2023-12-28 14:42:34 -08:00
Timothy Carambat
24227e48a7
Add LLM support for Google Gemini-Pro ( #492 )
...
resolves #489
2023-12-27 17:08:03 -08:00