fix: fix formatting issues

This commit is contained in:
chenzihong
2024-12-31 01:33:14 +08:00
parent 97bc42d896
commit 648645ef45

View File

@@ -47,7 +47,9 @@ os.environ["TOKENIZERS_PARALLELISM"] = "false"
@retry( @retry(
stop=stop_after_attempt(3), stop=stop_after_attempt(3),
wait=wait_exponential(multiplier=1, min=4, max=10), wait=wait_exponential(multiplier=1, min=4, max=10),
retry=retry_if_exception_type((RateLimitError, APIConnectionError, APITimeoutError)), retry=retry_if_exception_type(
(RateLimitError, APIConnectionError, APITimeoutError)
),
) )
async def openai_complete_if_cache( async def openai_complete_if_cache(
model, model,
@@ -108,7 +110,9 @@ async def openai_complete_if_cache(
@retry( @retry(
stop=stop_after_attempt(3), stop=stop_after_attempt(3),
wait=wait_exponential(multiplier=1, min=4, max=10), wait=wait_exponential(multiplier=1, min=4, max=10),
retry=retry_if_exception_type((RateLimitError, APIConnectionError, APIConnectionError)), retry=retry_if_exception_type(
(RateLimitError, APIConnectionError, APIConnectionError)
),
) )
async def azure_openai_complete_if_cache( async def azure_openai_complete_if_cache(
model, model,
@@ -259,7 +263,9 @@ def initialize_hf_model(model_name):
@retry( @retry(
stop=stop_after_attempt(3), stop=stop_after_attempt(3),
wait=wait_exponential(multiplier=1, min=4, max=10), wait=wait_exponential(multiplier=1, min=4, max=10),
retry=retry_if_exception_type((RateLimitError, APIConnectionError, APITimeoutError)), retry=retry_if_exception_type(
(RateLimitError, APIConnectionError, APITimeoutError)
),
) )
async def hf_model_if_cache( async def hf_model_if_cache(
model, model,
@@ -326,7 +332,9 @@ async def hf_model_if_cache(
@retry( @retry(
stop=stop_after_attempt(3), stop=stop_after_attempt(3),
wait=wait_exponential(multiplier=1, min=4, max=10), wait=wait_exponential(multiplier=1, min=4, max=10),
retry=retry_if_exception_type((RateLimitError, APIConnectionError, APITimeoutError)), retry=retry_if_exception_type(
(RateLimitError, APIConnectionError, APITimeoutError)
),
) )
async def ollama_model_if_cache( async def ollama_model_if_cache(
model, model,
@@ -444,7 +452,9 @@ def initialize_lmdeploy_pipeline(
@retry( @retry(
stop=stop_after_attempt(3), stop=stop_after_attempt(3),
wait=wait_exponential(multiplier=1, min=4, max=10), wait=wait_exponential(multiplier=1, min=4, max=10),
retry=retry_if_exception_type((RateLimitError, APIConnectionError, APITimeoutError)), retry=retry_if_exception_type(
(RateLimitError, APIConnectionError, APITimeoutError)
),
) )
async def lmdeploy_model_if_cache( async def lmdeploy_model_if_cache(
model, model,
@@ -704,7 +714,9 @@ async def lollms_model_complete(
@retry( @retry(
stop=stop_after_attempt(3), stop=stop_after_attempt(3),
wait=wait_exponential(multiplier=1, min=4, max=10), wait=wait_exponential(multiplier=1, min=4, max=10),
retry=retry_if_exception_type((RateLimitError, APIConnectionError, APITimeoutError)), retry=retry_if_exception_type(
(RateLimitError, APIConnectionError, APITimeoutError)
),
) )
async def zhipu_complete_if_cache( async def zhipu_complete_if_cache(
prompt: Union[str, List[Dict[str, str]]], prompt: Union[str, List[Dict[str, str]]],
@@ -834,7 +846,9 @@ async def zhipu_complete(
@retry( @retry(
stop=stop_after_attempt(3), stop=stop_after_attempt(3),
wait=wait_exponential(multiplier=1, min=4, max=60), wait=wait_exponential(multiplier=1, min=4, max=60),
retry=retry_if_exception_type((RateLimitError, APIConnectionError, APITimeoutError)), retry=retry_if_exception_type(
(RateLimitError, APIConnectionError, APITimeoutError)
),
) )
async def zhipu_embedding( async def zhipu_embedding(
texts: list[str], model: str = "embedding-3", api_key: str = None, **kwargs texts: list[str], model: str = "embedding-3", api_key: str = None, **kwargs
@@ -870,7 +884,9 @@ async def zhipu_embedding(
@retry( @retry(
stop=stop_after_attempt(3), stop=stop_after_attempt(3),
wait=wait_exponential(multiplier=1, min=4, max=60), wait=wait_exponential(multiplier=1, min=4, max=60),
retry=retry_if_exception_type((RateLimitError, APIConnectionError, APITimeoutError)), retry=retry_if_exception_type(
(RateLimitError, APIConnectionError, APITimeoutError)
),
) )
async def openai_embedding( async def openai_embedding(
texts: list[str], texts: list[str],
@@ -928,7 +944,9 @@ async def jina_embedding(
@retry( @retry(
stop=stop_after_attempt(3), stop=stop_after_attempt(3),
wait=wait_exponential(multiplier=1, min=4, max=60), wait=wait_exponential(multiplier=1, min=4, max=60),
retry=retry_if_exception_type((RateLimitError, APIConnectionError, APITimeoutError)), retry=retry_if_exception_type(
(RateLimitError, APIConnectionError, APITimeoutError)
),
) )
async def nvidia_openai_embedding( async def nvidia_openai_embedding(
texts: list[str], texts: list[str],
@@ -959,7 +977,9 @@ async def nvidia_openai_embedding(
@retry( @retry(
stop=stop_after_attempt(3), stop=stop_after_attempt(3),
wait=wait_exponential(multiplier=1, min=4, max=10), wait=wait_exponential(multiplier=1, min=4, max=10),
retry=retry_if_exception_type((RateLimitError, APIConnectionError, APITimeoutError)), retry=retry_if_exception_type(
(RateLimitError, APIConnectionError, APITimeoutError)
),
) )
async def azure_openai_embedding( async def azure_openai_embedding(
texts: list[str], texts: list[str],
@@ -990,7 +1010,9 @@ async def azure_openai_embedding(
@retry( @retry(
stop=stop_after_attempt(3), stop=stop_after_attempt(3),
wait=wait_exponential(multiplier=1, min=4, max=60), wait=wait_exponential(multiplier=1, min=4, max=60),
retry=retry_if_exception_type((RateLimitError, APIConnectionError, APITimeoutError)), retry=retry_if_exception_type(
(RateLimitError, APIConnectionError, APITimeoutError)
),
) )
async def siliconcloud_embedding( async def siliconcloud_embedding(
texts: list[str], texts: list[str],