Fix LLM binding variable name in create_app function
Corrected variable name from llm_binding_host Updated conditional checks for LLM bindings
This commit is contained in:
@@ -719,14 +719,14 @@ def create_app(args):
|
|||||||
|
|
||||||
# Create working directory if it doesn't exist
|
# Create working directory if it doesn't exist
|
||||||
Path(args.working_dir).mkdir(parents=True, exist_ok=True)
|
Path(args.working_dir).mkdir(parents=True, exist_ok=True)
|
||||||
if args.llm_binding_host == "lollms" or args.embedding_binding == "lollms":
|
if args.llm_binding == "lollms" or args.embedding_binding == "lollms":
|
||||||
from lightrag.llm.lollms import lollms_model_complete, lollms_embed
|
from lightrag.llm.lollms import lollms_model_complete, lollms_embed
|
||||||
if args.llm_binding_host == "ollama" or args.embedding_binding == "ollama":
|
if args.llm_binding == "ollama" or args.embedding_binding == "ollama":
|
||||||
from lightrag.llm.ollama import ollama_model_complete, ollama_embed
|
from lightrag.llm.ollama import ollama_model_complete, ollama_embed
|
||||||
if args.llm_binding_host == "openai" or args.embedding_binding == "openai":
|
if args.llm_binding == "openai" or args.embedding_binding == "openai":
|
||||||
from lightrag.llm.openai import openai_complete_if_cache, openai_embed
|
from lightrag.llm.openai import openai_complete_if_cache, openai_embed
|
||||||
if (
|
if (
|
||||||
args.llm_binding_host == "azure_openai"
|
args.llm_binding == "azure_openai"
|
||||||
or args.embedding_binding == "azure_openai"
|
or args.embedding_binding == "azure_openai"
|
||||||
):
|
):
|
||||||
from lightrag.llm.azure_openai import (
|
from lightrag.llm.azure_openai import (
|
||||||
|
Reference in New Issue
Block a user