From 588f35b6e9be04e36fdcce55fdfcf55ac2573355 Mon Sep 17 00:00:00 2001 From: sabaimran Date: Tue, 26 Sep 2023 10:57:35 -0700 Subject: [PATCH] Add max prompt size for gpt-3.5-turbo-16k --- src/khoj/processor/conversation/utils.py | 7 ++++++- 1 file changed, 6 insertions(+), 1 deletion(-) diff --git a/src/khoj/processor/conversation/utils.py b/src/khoj/processor/conversation/utils.py index 7bcac2d8..4a92c367 100644 --- a/src/khoj/processor/conversation/utils.py +++ b/src/khoj/processor/conversation/utils.py @@ -14,7 +14,12 @@ import queue from khoj.utils.helpers import merge_dicts logger = logging.getLogger(__name__) -max_prompt_size = {"gpt-3.5-turbo": 4096, "gpt-4": 8192, "llama-2-7b-chat.ggmlv3.q4_K_S.bin": 1548} +max_prompt_size = { + "gpt-3.5-turbo": 4096, + "gpt-4": 8192, + "llama-2-7b-chat.ggmlv3.q4_K_S.bin": 1548, + "gpt-3.5-turbo-16k": 15000, +} tokenizer = {"llama-2-7b-chat.ggmlv3.q4_K_S.bin": "hf-internal-testing/llama-tokenizer"}