Commit graph

30 commits

Author SHA1 Message Date
Timothy Carambat
c4f75feb08
Support historical message image inputs/attachments for n+1 queries ()
* Support historical message image inputs/attachments for n+1 queries

* patch gemini

* OpenRouter vision support cleanup

* xai vision history support

* Mistral logging

---------

Co-authored-by: shatfield4 <seanhatfield5@gmail.com>
2025-01-16 13:49:06 -08:00
Timothy Carambat
21af81085a
Add caching to Gemini /models ()
rename file typo
2025-01-13 13:12:03 -08:00
Timothy Carambat
b082c8e441
Add support for gemini authenticated models endpoint ()
* Add support for gemini authenticated models endpoint
add customModels entry
add un-authed fallback to default listing
separate models by expiermental status
resolves 

* add back improved logic for apiVersion decision making
2024-12-17 15:20:26 -08:00
Timothy Carambat
dd7c4675d3
LLM performance metric tracking ()
* WIP performance metric tracking

* fix: patch UI trying to .toFixed() null metric
Anthropic tracking migraiton
cleanup logs

* Apipie implmentation, not tested

* Cleanup Anthropic notes, Add support for AzureOpenAI tracking

* bedrock token metric tracking

* Cohere support

* feat: improve default stream handler to track for provider who are actually OpenAI compliant in usage reporting
add deepseek support

* feat: Add FireworksAI tracking reporting
fix: improve handler when usage:null is reported (why?)

* Add token reporting for GenericOpenAI

* token reporting for koboldcpp + lmstudio

* lint

* support Groq token tracking

* HF token tracking

* token tracking for togetherai

* LiteLLM token tracking

* linting + Mitral token tracking support

* XAI token metric reporting

* native provider runner

* LocalAI token tracking

* Novita token tracking

* OpenRouter token tracking

* Apipie stream metrics

* textwebgenui token tracking

* perplexity token reporting

* ollama token reporting

* lint

* put back comment

* Rip out LC ollama wrapper and use official library

* patch images with new ollama lib

* improve ollama offline message

* fix image handling in ollama llm provider

* lint

* NVIDIA NIM token tracking

* update openai compatbility responses

* UI/UX show/hide metrics on click for user preference

* update bedrock client

---------

Co-authored-by: shatfield4 <seanhatfield5@gmail.com>
2024-12-16 14:31:17 -08:00
timothycarambat
f8e91e1ffa patch gemini-2.0-key 2024-12-11 16:52:31 -08:00
timothycarambat
69b672b625 add gemini 1206 and gemini-2.0-flash exp models
connect 
2024-12-11 09:04:29 -08:00
timothycarambat
62be0cd0c5 add gemini-exp-1121 to expiermental set 2024-11-22 09:36:44 -08:00
timothycarambat
246152c024 Add gemini-exp-1121
resolves 
2024-11-21 11:02:43 -08:00
Timothy Carambat
26e2d8cc3b
Add more expiermental models from Gemini () 2024-11-20 09:52:33 -08:00
Timothy Carambat
c612239ecb
Add Gemini exp models ()
Add Gemini  models
resolves 
2024-09-11 13:03:14 -07:00
Timothy Carambat
99f2c25b1c
Agent Context window + context window refactor. ()
* Enable agent context windows to be accurate per provider:model

* Refactor model mapping to external file
Add token count to document length instead of char-count
refernce promptWindowLimit from AIProvider in central location

* remove unused imports
2024-08-15 12:13:28 -07:00
Timothy Carambat
f06ef6180d
add exp model to v1Beta () 2024-08-09 14:19:49 -07:00
RahSwe
c55ef33fce
Gemini Pro 1.5, API support for 2M context and new experimental model () 2024-08-02 10:24:31 -07:00
Timothy Carambat
38fc181238
Add multimodality support ()
* Add multimodality support

* Add Bedrock, KoboldCpp,LocalAI,and TextWebGenUI multi-modal

* temp dev build

* patch bad import

* noscrolls for windows dnd

* noscrolls for windows dnd

* update README

* update README

* add multimodal check
2024-07-31 10:47:49 -07:00
Timothy Carambat
0b845fbb1c
Deprecate .isSafe moderation ()
Add type defs to helpers
2024-06-28 15:32:30 -07:00
Sean Hatfield
3f78ef413b
[FEAT] Support for gemini-1.0-pro model and fixes to prompt window limit ()
support for gemini-1.0-pro model and fixes to prompt window limit
2024-05-29 08:17:35 +08:00
Sean Hatfield
cc7e7fb3ac
[FEAT] Add support for gemini-1.5-flash-latest model ()
* add support for gemini-1.5-flash-latest

* update comment in gemini LLM provider
2024-05-23 09:42:30 -07:00
Timothy Carambat
28eba636e9
Allow setting of safety thresholds for Gemini ()
* Allow setting of safety thresholds for Gemini

* linting
2024-05-20 13:17:00 -05:00
Timothy Carambat
01cf2fed17
Make native embedder the fallback for all LLMs () 2024-05-16 17:25:05 -07:00
Sean Hatfield
948ac8a3dd
[FIX] Validate messages schema for gemini provider ()
validate messages schema for gemini provider
2024-05-10 17:33:25 -07:00
Sean Hatfield
9feaad79cc
[CHORE] Remove sendChat and streamChat in all LLM providers ()
* remove sendChat and streamChat functions/references in all LLM providers

* remove unused imports

---------

Co-authored-by: timothycarambat <rambat1010@gmail.com>
2024-05-01 16:52:28 -07:00
Timothy Carambat
58b744771f
Add support for Gemini-1.5 Pro ()
* Add support for Gemini-1.5 Pro
bump @google/generative-ai pkg
Toggle apiVersion if beta model selected
resolves 

* update API messages due to package change
2024-04-19 08:59:46 -07:00
Timothy Carambat
0dd6001fa6
Patch Gemini/Google AI errors () 2024-03-26 17:20:12 -07:00
Timothy Carambat
0e46a11cb6
Stop generation button during stream-response ()
* Stop generation button during stream-response

* add custom stop icon

* add stop to thread chats
2024-03-12 15:21:27 -07:00
Timothy Carambat
c59ab9da0a
Refactor LLM chat backend ()
* refactor stream/chat/embed-stram to be a single execution logic path so that it is easier to maintain and build upon

* no thread in sync chat since only api uses it
adjust import locations
2024-02-14 12:32:07 -08:00
Timothy Carambat
aca5940650
Refactor handleStream to LLM Classes () 2024-02-07 08:15:14 -08:00
Sean Hatfield
c2c8fe9756
add support for mistral api ()
* add support for mistral api

* update docs to show support for Mistral

* add default temp to all providers, suggest different results per provider

---------

Co-authored-by: timothycarambat <rambat1010@gmail.com>
2024-01-17 14:42:05 -08:00
Sean Hatfield
90df37582b
Per workspace model selection ()
* WIP model selection per workspace (migrations and openai saves properly

* revert OpenAiOption

* add support for models per workspace for anthropic, localAi, ollama, openAi, and togetherAi

* remove unneeded comments

* update logic for when LLMProvider is reset, reset Ai provider files with master

* remove frontend/api reset of workspace chat and move logic to updateENV
add postUpdate callbacks to envs

* set preferred model for chat on class instantiation

* remove extra param

* linting

* remove unused var

* refactor chat model selection on workspace

* linting

* add fallback for base path to localai models

---------

Co-authored-by: timothycarambat <rambat1010@gmail.com>
2024-01-17 12:59:25 -08:00
Timothy Carambat
6d5968bf7e
Llm chore cleanup ()
* move internal functions to private in class
simplify lc message convertor

* Fix hanging Context text when none is present
2023-12-28 14:42:34 -08:00
Timothy Carambat
24227e48a7
Add LLM support for Google Gemini-Pro ()
resolves 
2023-12-27 17:08:03 -08:00