From 7c4bbe2474a251f1e17e77f54436637303b80311 Mon Sep 17 00:00:00 2001 From: yuanxiaobin Date: Fri, 6 Dec 2024 10:40:48 +0800 Subject: [PATCH] =?UTF-8?q?=E4=BF=AE=E5=A4=8D=20args=5Fhash=E5=9C=A8?= =?UTF-8?q?=E4=BD=BF=E7=94=A8=E5=B8=B8=E8=A7=84=E7=BC=93=E5=AD=98=E6=97=B6?= =?UTF-8?q?=E5=80=99=E6=89=8D=E8=AE=A1=E7=AE=97=E5=AF=BC=E8=87=B4embedding?= =?UTF-8?q?=E7=BC=93=E5=AD=98=E6=97=B6=E6=B2=A1=E6=9C=89=E8=AE=A1=E7=AE=97?= =?UTF-8?q?=E7=9A=84bug?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- lightrag/llm.py | 4 ---- 1 file changed, 4 deletions(-) diff --git a/lightrag/llm.py b/lightrag/llm.py index 97c903d2..fef8c9a3 100644 --- a/lightrag/llm.py +++ b/lightrag/llm.py @@ -162,7 +162,6 @@ async def azure_openai_complete_if_cache( if prompt is not None: messages.append({"role": "user", "content": prompt}) - hashing_kv: BaseKVStorage = kwargs.pop("hashing_kv", None) if hashing_kv is not None: # Calculate args_hash only when using cache args_hash = compute_args_hash(model, messages) @@ -373,7 +372,6 @@ async def hf_model_if_cache( messages.extend(history_messages) messages.append({"role": "user", "content": prompt}) - hashing_kv: BaseKVStorage = kwargs.pop("hashing_kv", None) if hashing_kv is not None: # Calculate args_hash only when using cache args_hash = compute_args_hash(model, messages) @@ -491,7 +489,6 @@ async def ollama_model_if_cache( messages.extend(history_messages) messages.append({"role": "user", "content": prompt}) - hashing_kv: BaseKVStorage = kwargs.pop("hashing_kv", None) if hashing_kv is not None: # Calculate args_hash only when using cache args_hash = compute_args_hash(model, messages) @@ -652,7 +649,6 @@ async def lmdeploy_model_if_cache( messages.extend(history_messages) messages.append({"role": "user", "content": prompt}) - hashing_kv: BaseKVStorage = kwargs.pop("hashing_kv", None) if hashing_kv is not None: # Calculate args_hash only when using cache args_hash = compute_args_hash(model, messages)