From 23cabbe7a35ee7fe6cd16504e76a662ddcdce315 Mon Sep 17 00:00:00 2001 From: Ahmad Hatahet Date: Sat, 30 Nov 2024 17:16:07 +0100 Subject: [PATCH] update max_token_size according to openai doc --- lightrag/llm.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/lightrag/llm.py b/lightrag/llm.py index 3c4f9543..19d22bc8 100644 --- a/lightrag/llm.py +++ b/lightrag/llm.py @@ -556,7 +556,7 @@ async def openai_embedding( return np.array([dp.embedding for dp in response.data]) -@wrap_embedding_func_with_attrs(embedding_dim=1536, max_token_size=8192) +@wrap_embedding_func_with_attrs(embedding_dim=1536, max_token_size=8191) @retry( stop=stop_after_attempt(3), wait=wait_exponential(multiplier=1, min=4, max=10),