Timothy Carambat
c4f75feb08
Support historical message image inputs/attachments for n+1 queries ( #2919 )
...
* Support historical message image inputs/attachments for n+1 queries
* patch gemini
* OpenRouter vision support cleanup
* xai vision history support
* Mistral logging
---------
Co-authored-by: shatfield4 <seanhatfield5@gmail.com>
2025-01-16 13:49:06 -08:00
timothycarambat
4b2bb529c9
enable leftover mlock setting
2024-12-28 17:48:24 -08:00
Timothy Carambat
a51de73aaa
update ollama performance mode ( #2874 )
2024-12-18 11:21:35 -08:00
Timothy Carambat
dd7c4675d3
LLM performance metric tracking ( #2825 )
...
* WIP performance metric tracking
* fix: patch UI trying to .toFixed() null metric
Anthropic tracking migraiton
cleanup logs
* Apipie implmentation, not tested
* Cleanup Anthropic notes, Add support for AzureOpenAI tracking
* bedrock token metric tracking
* Cohere support
* feat: improve default stream handler to track for provider who are actually OpenAI compliant in usage reporting
add deepseek support
* feat: Add FireworksAI tracking reporting
fix: improve handler when usage:null is reported (why?)
* Add token reporting for GenericOpenAI
* token reporting for koboldcpp + lmstudio
* lint
* support Groq token tracking
* HF token tracking
* token tracking for togetherai
* LiteLLM token tracking
* linting + Mitral token tracking support
* XAI token metric reporting
* native provider runner
* LocalAI token tracking
* Novita token tracking
* OpenRouter token tracking
* Apipie stream metrics
* textwebgenui token tracking
* perplexity token reporting
* ollama token reporting
* lint
* put back comment
* Rip out LC ollama wrapper and use official library
* patch images with new ollama lib
* improve ollama offline message
* fix image handling in ollama llm provider
* lint
* NVIDIA NIM token tracking
* update openai compatbility responses
* UI/UX show/hide metrics on click for user preference
* update bedrock client
---------
Co-authored-by: shatfield4 <seanhatfield5@gmail.com>
2024-12-16 14:31:17 -08:00
Timothy Carambat
99f2c25b1c
Agent Context window + context window refactor. ( #2126 )
...
* Enable agent context windows to be accurate per provider:model
* Refactor model mapping to external file
Add token count to document length instead of char-count
refernce promptWindowLimit from AIProvider in central location
* remove unused imports
2024-08-15 12:13:28 -07:00
Sean Hatfield
7273c892a1
Ollama performance mode option ( #2014 )
...
* ollama performance mode option
* Change ENV prop
Move perf setting to advanced
---------
Co-authored-by: timothycarambat <rambat1010@gmail.com>
2024-08-02 13:29:17 -07:00
Timothy Carambat
38fc181238
Add multimodality support ( #2001 )
...
* Add multimodality support
* Add Bedrock, KoboldCpp,LocalAI,and TextWebGenUI multi-modal
* temp dev build
* patch bad import
* noscrolls for windows dnd
* noscrolls for windows dnd
* update README
* update README
* add multimodal check
2024-07-31 10:47:49 -07:00
Timothy Carambat
76aa2a4fd4
Implement support for selecting basic keep_alive
times for Ollama ( #1920 )
2024-07-22 14:44:47 -07:00
Timothy Carambat
0b845fbb1c
Deprecate .isSafe
moderation ( #1790 )
...
Add type defs to helpers
2024-06-28 15:32:30 -07:00
Timothy Carambat
01cf2fed17
Make native embedder the fallback for all LLMs ( #1427 )
2024-05-16 17:25:05 -07:00
Sean Hatfield
9feaad79cc
[CHORE] Remove sendChat and streamChat in all LLM providers ( #1260 )
...
* remove sendChat and streamChat functions/references in all LLM providers
* remove unused imports
---------
Co-authored-by: timothycarambat <rambat1010@gmail.com>
2024-05-01 16:52:28 -07:00
Timothy Carambat
94017e2b51
bump langchain deps ( #1231 )
...
* bump langchain deps
* patch native and ollama providers remove deprecated deps
---------
Co-authored-by: shatfield4 <seanhatfield5@gmail.com>
2024-04-30 12:04:24 -07:00
Timothy Carambat
df2aac9f3c
useMLock for Ollama API chats ( #1014 )
2024-04-02 10:43:04 -07:00
Timothy Carambat
0e46a11cb6
Stop generation button during stream-response ( #892 )
...
* Stop generation button during stream-response
* add custom stop icon
* add stop to thread chats
2024-03-12 15:21:27 -07:00
Sean Hatfield
e99c74aec1
[DOCS] Update Docker documentation to show how to setup Ollama with Dockerized version of AnythingLLM ( #774 )
...
* update HOW_TO_USE_DOCKER to help with Ollama setup using docker
* update HOW_TO_USE_DOCKER
* styles update
* create separate README for ollama and link to it in HOW_TO_USE_DOCKER
* styling update
2024-02-21 18:42:32 -08:00
Timothy Carambat
c59ab9da0a
Refactor LLM chat backend ( #717 )
...
* refactor stream/chat/embed-stram to be a single execution logic path so that it is easier to maintain and build upon
* no thread in sync chat since only api uses it
adjust import locations
2024-02-14 12:32:07 -08:00
Timothy Carambat
f490c35456
Recover from fatal Ollama crash from LangChain library ( #693 )
...
Resolve fatal crash from Ollama failure
2024-02-07 16:23:17 -08:00
Timothy Carambat
aca5940650
Refactor handleStream to LLM Classes ( #685 )
2024-02-07 08:15:14 -08:00
Sean Hatfield
c2c8fe9756
add support for mistral api ( #610 )
...
* add support for mistral api
* update docs to show support for Mistral
* add default temp to all providers, suggest different results per provider
---------
Co-authored-by: timothycarambat <rambat1010@gmail.com>
2024-01-17 14:42:05 -08:00
Sean Hatfield
90df37582b
Per workspace model selection ( #582 )
...
* WIP model selection per workspace (migrations and openai saves properly
* revert OpenAiOption
* add support for models per workspace for anthropic, localAi, ollama, openAi, and togetherAi
* remove unneeded comments
* update logic for when LLMProvider is reset, reset Ai provider files with master
* remove frontend/api reset of workspace chat and move logic to updateENV
add postUpdate callbacks to envs
* set preferred model for chat on class instantiation
* remove extra param
* linting
* remove unused var
* refactor chat model selection on workspace
* linting
* add fallback for base path to localai models
---------
Co-authored-by: timothycarambat <rambat1010@gmail.com>
2024-01-17 12:59:25 -08:00
Timothy Carambat
6d5968bf7e
Llm chore cleanup ( #501 )
...
* move internal functions to private in class
simplify lc message convertor
* Fix hanging Context text when none is present
2023-12-28 14:42:34 -08:00
Timothy Carambat
2a1202de54
Patch Ollama Streaming chunk issues ( #500 )
...
Replace stream/sync chats with Langchain interface for now
connect #499
ref: https://github.com/Mintplex-Labs/anything-llm/issues/495#issuecomment-1871476091
2023-12-28 13:59:47 -08:00
Timothy Carambat
e0a0a8976d
Add Ollama as LLM provider option ( #494 )
...
* Add support for Ollama as LLM provider
resolves #493
2023-12-27 17:21:47 -08:00