Fix linting
This commit is contained in:
@@ -100,10 +100,10 @@ def create_openai_async_client(
|
|||||||
stop=stop_after_attempt(3),
|
stop=stop_after_attempt(3),
|
||||||
wait=wait_exponential(multiplier=1, min=4, max=10),
|
wait=wait_exponential(multiplier=1, min=4, max=10),
|
||||||
retry=(
|
retry=(
|
||||||
retry_if_exception_type(RateLimitError) |
|
retry_if_exception_type(RateLimitError)
|
||||||
retry_if_exception_type(APIConnectionError) |
|
| retry_if_exception_type(APIConnectionError)
|
||||||
retry_if_exception_type(APITimeoutError) |
|
| retry_if_exception_type(APITimeoutError)
|
||||||
retry_if_exception_type(InvalidResponseError)
|
| retry_if_exception_type(InvalidResponseError)
|
||||||
),
|
),
|
||||||
)
|
)
|
||||||
async def openai_complete_if_cache(
|
async def openai_complete_if_cache(
|
||||||
@@ -230,21 +230,33 @@ async def openai_complete_if_cache(
|
|||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.error(f"Error in stream response: {str(e)}")
|
logger.error(f"Error in stream response: {str(e)}")
|
||||||
# Try to clean up resources if possible
|
# Try to clean up resources if possible
|
||||||
if iteration_started and hasattr(response, "aclose") and callable(getattr(response, "aclose", None)):
|
if (
|
||||||
|
iteration_started
|
||||||
|
and hasattr(response, "aclose")
|
||||||
|
and callable(getattr(response, "aclose", None))
|
||||||
|
):
|
||||||
try:
|
try:
|
||||||
await response.aclose()
|
await response.aclose()
|
||||||
logger.debug("Successfully closed stream response after error")
|
logger.debug("Successfully closed stream response after error")
|
||||||
except Exception as close_error:
|
except Exception as close_error:
|
||||||
logger.warning(f"Failed to close stream response: {close_error}")
|
logger.warning(
|
||||||
|
f"Failed to close stream response: {close_error}"
|
||||||
|
)
|
||||||
raise
|
raise
|
||||||
finally:
|
finally:
|
||||||
# Ensure resources are released even if no exception occurs
|
# Ensure resources are released even if no exception occurs
|
||||||
if iteration_started and hasattr(response, "aclose") and callable(getattr(response, "aclose", None)):
|
if (
|
||||||
|
iteration_started
|
||||||
|
and hasattr(response, "aclose")
|
||||||
|
and callable(getattr(response, "aclose", None))
|
||||||
|
):
|
||||||
try:
|
try:
|
||||||
await response.aclose()
|
await response.aclose()
|
||||||
logger.debug("Successfully closed stream response")
|
logger.debug("Successfully closed stream response")
|
||||||
except Exception as close_error:
|
except Exception as close_error:
|
||||||
logger.warning(f"Failed to close stream response in finally block: {close_error}")
|
logger.warning(
|
||||||
|
f"Failed to close stream response in finally block: {close_error}"
|
||||||
|
)
|
||||||
|
|
||||||
return inner()
|
return inner()
|
||||||
|
|
||||||
@@ -373,9 +385,9 @@ async def nvidia_openai_complete(
|
|||||||
stop=stop_after_attempt(3),
|
stop=stop_after_attempt(3),
|
||||||
wait=wait_exponential(multiplier=1, min=4, max=60),
|
wait=wait_exponential(multiplier=1, min=4, max=60),
|
||||||
retry=(
|
retry=(
|
||||||
retry_if_exception_type(RateLimitError) |
|
retry_if_exception_type(RateLimitError)
|
||||||
retry_if_exception_type(APIConnectionError) |
|
| retry_if_exception_type(APIConnectionError)
|
||||||
retry_if_exception_type(APITimeoutError)
|
| retry_if_exception_type(APITimeoutError)
|
||||||
),
|
),
|
||||||
)
|
)
|
||||||
async def openai_embed(
|
async def openai_embed(
|
||||||
|
Reference in New Issue
Block a user