Removed verbose module-level documentation
This commit is contained in:
@@ -1,38 +1,3 @@
|
|||||||
"""
|
|
||||||
LlamaIndex LLM Interface Module
|
|
||||||
==========================
|
|
||||||
|
|
||||||
This module provides interfaces for interacting with LlamaIndex's language models,
|
|
||||||
including text generation and embedding capabilities.
|
|
||||||
|
|
||||||
Author: Lightrag team
|
|
||||||
Created: 2024-03-19
|
|
||||||
License: MIT License
|
|
||||||
|
|
||||||
Version: 1.0.0
|
|
||||||
|
|
||||||
Change Log:
|
|
||||||
- 1.0.0 (2024-03-19): Initial release
|
|
||||||
* Added async chat completion support
|
|
||||||
* Added embedding generation
|
|
||||||
* Added stream response capability
|
|
||||||
* Added support for external settings configuration
|
|
||||||
* Added OpenAI-style message handling
|
|
||||||
|
|
||||||
Dependencies:
|
|
||||||
- llama_index
|
|
||||||
- numpy
|
|
||||||
- pipmaster
|
|
||||||
- Python >= 3.10
|
|
||||||
|
|
||||||
Usage:
|
|
||||||
from lightrag.llm.llama_index_impl import llama_index_complete, llama_index_embed
|
|
||||||
"""
|
|
||||||
|
|
||||||
__version__ = "1.0.0"
|
|
||||||
__author__ = "lightrag Team"
|
|
||||||
__status__ = "Production"
|
|
||||||
|
|
||||||
import pipmaster as pm
|
import pipmaster as pm
|
||||||
from core.logging_config import setup_logger
|
from core.logging_config import setup_logger
|
||||||
from llama_index.core.llms import (
|
from llama_index.core.llms import (
|
||||||
|
Reference in New Issue
Block a user