Add tokenizer to global_config
This commit is contained in:
@@ -322,11 +322,6 @@ class LightRAG:
|
|||||||
**self.vector_db_storage_cls_kwargs,
|
**self.vector_db_storage_cls_kwargs,
|
||||||
}
|
}
|
||||||
|
|
||||||
# Show config
|
|
||||||
global_config = asdict(self)
|
|
||||||
_print_config = ",\n ".join([f"{k} = {v}" for k, v in global_config.items()])
|
|
||||||
logger.debug(f"LightRAG init with param:\n {_print_config}\n")
|
|
||||||
|
|
||||||
# Init Tokenizer
|
# Init Tokenizer
|
||||||
# Post-initialization hook to handle backward compatabile tokenizer initialization based on provided parameters
|
# Post-initialization hook to handle backward compatabile tokenizer initialization based on provided parameters
|
||||||
if self.tokenizer is None:
|
if self.tokenizer is None:
|
||||||
@@ -335,6 +330,11 @@ class LightRAG:
|
|||||||
else:
|
else:
|
||||||
self.tokenizer = TiktokenTokenizer()
|
self.tokenizer = TiktokenTokenizer()
|
||||||
|
|
||||||
|
# Fix global_config now
|
||||||
|
global_config = asdict(self)
|
||||||
|
_print_config = ",\n ".join([f"{k} = {v}" for k, v in global_config.items()])
|
||||||
|
logger.debug(f"LightRAG init with param:\n {_print_config}\n")
|
||||||
|
|
||||||
# Init Embedding
|
# Init Embedding
|
||||||
self.embedding_func = limit_async_func_call(self.embedding_func_max_async)( # type: ignore
|
self.embedding_func = limit_async_func_call(self.embedding_func_max_async)( # type: ignore
|
||||||
self.embedding_func
|
self.embedding_func
|
||||||
|
Reference in New Issue
Block a user