Update default offline models enabled

This commit is contained in:
Debanjum 2024-11-18 16:22:32 -08:00
parent 45c623f95c
commit 817601872f
2 changed files with 8 additions and 6 deletions

View file

@ -49,8 +49,6 @@ except ImportError:
model_to_prompt_size = { model_to_prompt_size = {
# OpenAI Models # OpenAI Models
"gpt-3.5-turbo": 12000,
"gpt-4-turbo-preview": 20000,
"gpt-4o": 20000, "gpt-4o": 20000,
"gpt-4o-mini": 20000, "gpt-4o-mini": 20000,
"o1-preview": 20000, "o1-preview": 20000,
@ -60,11 +58,14 @@ model_to_prompt_size = {
"gemini-1.5-pro": 20000, "gemini-1.5-pro": 20000,
# Anthropic Models # Anthropic Models
"claude-3-5-sonnet-20240620": 20000, "claude-3-5-sonnet-20240620": 20000,
"claude-3-opus-20240229": 20000, "claude-3-5-haiku-20241022": 20000,
# Offline Models # Offline Models
"TheBloke/Mistral-7B-Instruct-v0.2-GGUF": 3500,
"NousResearch/Hermes-2-Pro-Mistral-7B-GGUF": 3500,
"bartowski/Meta-Llama-3.1-8B-Instruct-GGUF": 20000, "bartowski/Meta-Llama-3.1-8B-Instruct-GGUF": 20000,
"bartowski/Meta-Llama-3.1-8B-Instruct-GGUF": 20000,
"bartowski/Llama-3.2-3B-Instruct-GGUF": 20000,
"bartowski/gemma-2-9b-it-GGUF": 6000,
"bartowski/gemma-2-2b-it-GGUF": 6000,
"Qwen/Qwen2.5-14B-Instruct-GGUF": 20000,
} }
model_to_tokenizer: Dict[str, str] = {} model_to_tokenizer: Dict[str, str] = {}

View file

@ -10,9 +10,10 @@ telemetry_server = "https://khoj.beta.haletic.com/v1/telemetry"
content_directory = "~/.khoj/content/" content_directory = "~/.khoj/content/"
default_offline_chat_models = [ default_offline_chat_models = [
"bartowski/Meta-Llama-3.1-8B-Instruct-GGUF", "bartowski/Meta-Llama-3.1-8B-Instruct-GGUF",
"bartowski/Llama-3.2-3B-Instruct-GGUF",
"bartowski/gemma-2-9b-it-GGUF", "bartowski/gemma-2-9b-it-GGUF",
"bartowski/gemma-2-2b-it-GGUF", "bartowski/gemma-2-2b-it-GGUF",
"bartowski/Phi-3.5-mini-instruct-GGUF", "Qwen/Qwen2.5-14B-Instruct-GGUF",
] ]
default_openai_chat_models = ["gpt-4o-mini", "gpt-4o"] default_openai_chat_models = ["gpt-4o-mini", "gpt-4o"]
default_gemini_chat_models = ["gemini-1.5-flash", "gemini-1.5-pro"] default_gemini_chat_models = ["gemini-1.5-flash", "gemini-1.5-pro"]