inference running locally. use neo4j next
This commit is contained in:
@@ -72,7 +72,9 @@ async def openai_complete_if_cache(
|
||||
|
||||
@retry(
|
||||
stop=stop_after_attempt(3),
|
||||
wait=wait_exponential(multiplier=1, min=4, max=10),
|
||||
#kw_
|
||||
wait=wait_exponential(multiplier=1, min=4, max=60),
|
||||
# wait=wait_exponential(multiplier=1, min=4, max=10),
|
||||
retry=retry_if_exception_type((RateLimitError, APIConnectionError, Timeout)),
|
||||
)
|
||||
async def azure_openai_complete_if_cache(model,
|
||||
|
Reference in New Issue
Block a user