Merge pull request #771 from enriquecatala/fix-litellm-litellm-proxy-support
Enable LiteLLM proxy with embedding_binding_host
This commit is contained in:
@@ -1027,7 +1027,8 @@ def create_app(args):
|
|||||||
if args.embedding_binding == "azure_openai"
|
if args.embedding_binding == "azure_openai"
|
||||||
else openai_embed(
|
else openai_embed(
|
||||||
texts,
|
texts,
|
||||||
model=args.embedding_model, # no host is used for openai,
|
model=args.embedding_model,
|
||||||
|
base_url=args.embedding_binding_host,
|
||||||
api_key=args.embedding_binding_api_key,
|
api_key=args.embedding_binding_api_key,
|
||||||
),
|
),
|
||||||
)
|
)
|
||||||
|
Reference in New Issue
Block a user