From b7cce9312f8734646be0459e0a21afd815015272 Mon Sep 17 00:00:00 2001 From: yangdx Date: Mon, 17 Feb 2025 12:34:54 +0800 Subject: [PATCH] Fix linting --- lightrag/llm/openai.py | 5 ++--- 1 file changed, 2 insertions(+), 3 deletions(-) diff --git a/lightrag/llm/openai.py b/lightrag/llm/openai.py index 96255086..024f7f52 100644 --- a/lightrag/llm/openai.py +++ b/lightrag/llm/openai.py @@ -115,7 +115,7 @@ async def openai_complete_if_cache( # Set openai logger level to INFO when VERBOSE_DEBUG is off if not VERBOSE_DEBUG and logger.level == logging.DEBUG: logging.getLogger("openai").setLevel(logging.INFO) - + openai_async_client = ( AsyncOpenAI(default_headers=default_headers, api_key=api_key) if base_url is None @@ -136,9 +136,8 @@ async def openai_complete_if_cache( logger.debug(f"Additional kwargs: {kwargs}") verbose_debug(f"Query: {prompt}") verbose_debug(f"System prompt: {system_prompt}") - # logger.debug(f"Messages: {messages}") - try: + try: if "response_format" in kwargs: response = await openai_async_client.beta.chat.completions.parse( model=model, messages=messages, **kwargs