From ae37454c4243ee6ef205f88d349667ca8ab5bd1e Mon Sep 17 00:00:00 2001 From: yangdx Date: Mon, 17 Feb 2025 12:28:49 +0800 Subject: [PATCH] Replace verbose_debug with logger.debug for token logging. - Removed unused verbose_debug import - Updated debug logging in kg_query - Updated debug logging in mix_kg_vector_query - Updated debug logging in kg_query_with_keywords --- lightrag/operate.py | 7 +++---- 1 file changed, 3 insertions(+), 4 deletions(-) diff --git a/lightrag/operate.py b/lightrag/operate.py index 23764957..cc5dffe7 100644 --- a/lightrag/operate.py +++ b/lightrag/operate.py @@ -24,7 +24,6 @@ from .utils import ( CacheData, statistic_data, get_conversation_turns, - verbose_debug, ) from .base import ( BaseGraphStorage, @@ -689,7 +688,7 @@ async def kg_query( return sys_prompt len_of_prompts = len(encode_string_by_tiktoken(query + sys_prompt)) - verbose_debug(f"[kg_query]Prompt Tokens: {len_of_prompts}") + logger.debug(f"[kg_query]Prompt Tokens: {len_of_prompts}") response = await use_model_func( query, @@ -978,7 +977,7 @@ async def mix_kg_vector_query( return sys_prompt len_of_prompts = len(encode_string_by_tiktoken(query + sys_prompt)) - verbose_debug(f"[mix_kg_vector_query]Prompt Tokens: {len_of_prompts}") + logger.debug(f"[mix_kg_vector_query]Prompt Tokens: {len_of_prompts}") # 6. Generate response response = await use_model_func( @@ -1808,7 +1807,7 @@ async def kg_query_with_keywords( return sys_prompt len_of_prompts = len(encode_string_by_tiktoken(query + sys_prompt)) - verbose_debug(f"[kg_query_with_keywords]Prompt Tokens: {len_of_prompts}") + logger.debug(f"[kg_query_with_keywords]Prompt Tokens: {len_of_prompts}") response = await use_model_func( query,