From 991fa339864ba2e17254c85870e97cf80a627ac4 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Enrique=20Catal=C3=A1?= Date: Thu, 13 Feb 2025 22:45:03 +0100 Subject: [PATCH 1/3] Enable LiteLLM proxy with embedding_binding_host --- lightrag/api/lightrag_server.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/lightrag/api/lightrag_server.py b/lightrag/api/lightrag_server.py index fe52f592..4cc6f775 100644 --- a/lightrag/api/lightrag_server.py +++ b/lightrag/api/lightrag_server.py @@ -1021,7 +1021,8 @@ def create_app(args): if args.embedding_binding == "azure_openai" else openai_embed( texts, - model=args.embedding_model, # no host is used for openai, + model=args.embedding_model, + base_url=args.embedding_binding_host, # If you decide to use litellm as a proxy for azure openai, this is relevant api_key=args.embedding_binding_api_key, ), ) From 0ece8e102d5d39ff1a9d55eefabea4bc8abec295 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Enrique=20Catal=C3=A1?= Date: Fri, 14 Feb 2025 00:09:32 +0100 Subject: [PATCH 2/3] lint fix --- lightrag/api/lightrag_server.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/lightrag/api/lightrag_server.py b/lightrag/api/lightrag_server.py index 4cc6f775..b64c9efd 100644 --- a/lightrag/api/lightrag_server.py +++ b/lightrag/api/lightrag_server.py @@ -1021,8 +1021,8 @@ def create_app(args): if args.embedding_binding == "azure_openai" else openai_embed( texts, - model=args.embedding_model, - base_url=args.embedding_binding_host, # If you decide to use litellm as a proxy for azure openai, this is relevant + model=args.embedding_model, + base_url=args.embedding_binding_host, api_key=args.embedding_binding_api_key, ), ) From f9d69b889e101e0010c3d5a7b0a01bb84a1833e1 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Enrique=20Catal=C3=A1?= Date: Fri, 14 Feb 2025 00:16:59 +0100 Subject: [PATCH 3/3] ruff reformat --- lightrag/api/lightrag_server.py | 1 + 1 file changed, 1 insertion(+) diff --git a/lightrag/api/lightrag_server.py b/lightrag/api/lightrag_server.py index b64c9efd..cf426042 100644 --- a/lightrag/api/lightrag_server.py +++ b/lightrag/api/lightrag_server.py @@ -101,6 +101,7 @@ def estimate_tokens(text: str) -> int: return int(tokens) + def get_default_host(binding_type: str) -> str: default_hosts = { "ollama": os.getenv("LLM_BINDING_HOST", "http://localhost:11434"),