From 27a52ddf371679b62ba96c04d61f3fe0726fb39a Mon Sep 17 00:00:00 2001 From: Saifeddine ALOUI Date: Thu, 26 Dec 2024 23:39:10 +0100 Subject: [PATCH] fixed linting --- lightrag/api/azure_openai_lightrag_server.py | 12 ++++++++++-- lightrag/api/lollms_lightrag_server.py | 13 +++++++++++-- lightrag/api/ollama_lightrag_server.py | 13 +++++++++++-- lightrag/api/openai_lightrag_server.py | 13 +++++++++++-- 4 files changed, 43 insertions(+), 8 deletions(-) diff --git a/lightrag/api/azure_openai_lightrag_server.py b/lightrag/api/azure_openai_lightrag_server.py index 084f3caa..6b20caac 100644 --- a/lightrag/api/azure_openai_lightrag_server.py +++ b/lightrag/api/azure_openai_lightrag_server.py @@ -309,7 +309,11 @@ def create_app(args): try: response = await rag.aquery( request.query, - param=QueryParam(mode=request.mode, stream=False, only_need_context=request.only_need_context), + param=QueryParam( + mode=request.mode, + stream=False, + only_need_context=request.only_need_context, + ), ) return QueryResponse(response=response) except Exception as e: @@ -320,7 +324,11 @@ def create_app(args): try: response = await rag.aquery( request.query, - param=QueryParam(mode=request.mode, stream=True, only_need_context=request.only_need_context), + param=QueryParam( + mode=request.mode, + stream=True, + only_need_context=request.only_need_context, + ), ) if inspect.isasyncgen(response): diff --git a/lightrag/api/lollms_lightrag_server.py b/lightrag/api/lollms_lightrag_server.py index 45b4dc4b..1ce7b259 100644 --- a/lightrag/api/lollms_lightrag_server.py +++ b/lightrag/api/lollms_lightrag_server.py @@ -267,7 +267,11 @@ def create_app(args): try: response = await rag.aquery( request.query, - param=QueryParam(mode=request.mode, stream=request.stream, only_need_context=request.only_need_context), + param=QueryParam( + mode=request.mode, + stream=request.stream, + only_need_context=request.only_need_context, + ), ) if request.stream: @@ -284,7 +288,12 @@ def create_app(args): async def query_text_stream(request: QueryRequest): try: response = rag.query( - request.query, param=QueryParam(mode=request.mode, stream=True, only_need_context=request.only_need_context) + request.query, + param=QueryParam( + mode=request.mode, + stream=True, + only_need_context=request.only_need_context, + ), ) async def stream_generator(): diff --git a/lightrag/api/ollama_lightrag_server.py b/lightrag/api/ollama_lightrag_server.py index 5bbc32c2..40f617c6 100644 --- a/lightrag/api/ollama_lightrag_server.py +++ b/lightrag/api/ollama_lightrag_server.py @@ -267,7 +267,11 @@ def create_app(args): try: response = await rag.aquery( request.query, - param=QueryParam(mode=request.mode, stream=request.stream, only_need_context=request.only_need_context), + param=QueryParam( + mode=request.mode, + stream=request.stream, + only_need_context=request.only_need_context, + ), ) if request.stream: @@ -284,7 +288,12 @@ def create_app(args): async def query_text_stream(request: QueryRequest): try: response = rag.query( - request.query, param=QueryParam(mode=request.mode, stream=True, only_need_context=request.only_need_context) + request.query, + param=QueryParam( + mode=request.mode, + stream=True, + only_need_context=request.only_need_context, + ), ) async def stream_generator(): diff --git a/lightrag/api/openai_lightrag_server.py b/lightrag/api/openai_lightrag_server.py index 051bb1ff..119d6900 100644 --- a/lightrag/api/openai_lightrag_server.py +++ b/lightrag/api/openai_lightrag_server.py @@ -271,7 +271,11 @@ def create_app(args): try: response = await rag.aquery( request.query, - param=QueryParam(mode=request.mode, stream=request.stream, only_need_context=request.only_need_context), + param=QueryParam( + mode=request.mode, + stream=request.stream, + only_need_context=request.only_need_context, + ), ) if request.stream: @@ -288,7 +292,12 @@ def create_app(args): async def query_text_stream(request: QueryRequest): try: response = rag.query( - request.query, param=QueryParam(mode=request.mode, stream=True, only_need_context=request.only_need_context) + request.query, + param=QueryParam( + mode=request.mode, + stream=True, + only_need_context=request.only_need_context, + ), ) async def stream_generator():