From cd01ec64d348ecd09ffde84c2784b5f1a35811f3 Mon Sep 17 00:00:00 2001 From: yangdx Date: Sun, 20 Apr 2025 14:51:11 +0800 Subject: [PATCH] Add tokenizer to global_config --- lightrag/lightrag.py | 10 +++++----- 1 file changed, 5 insertions(+), 5 deletions(-) diff --git a/lightrag/lightrag.py b/lightrag/lightrag.py index 90a68aa9..2acde7e9 100644 --- a/lightrag/lightrag.py +++ b/lightrag/lightrag.py @@ -322,11 +322,6 @@ class LightRAG: **self.vector_db_storage_cls_kwargs, } - # Show config - global_config = asdict(self) - _print_config = ",\n ".join([f"{k} = {v}" for k, v in global_config.items()]) - logger.debug(f"LightRAG init with param:\n {_print_config}\n") - # Init Tokenizer # Post-initialization hook to handle backward compatabile tokenizer initialization based on provided parameters if self.tokenizer is None: @@ -335,6 +330,11 @@ class LightRAG: else: self.tokenizer = TiktokenTokenizer() + # Fix global_config now + global_config = asdict(self) + _print_config = ",\n ".join([f"{k} = {v}" for k, v in global_config.items()]) + logger.debug(f"LightRAG init with param:\n {_print_config}\n") + # Init Embedding self.embedding_func = limit_async_func_call(self.embedding_func_max_async)( # type: ignore self.embedding_func