diff --git a/lightrag/api/lightrag_ollama.py b/lightrag/api/lightrag_ollama.py index af3f22ee..82329806 100644 --- a/lightrag/api/lightrag_ollama.py +++ b/lightrag/api/lightrag_ollama.py @@ -243,6 +243,7 @@ class SearchMode(str, Enum): local = "local" global_ = "global" # 使用 global_ 因为 global 是 Python 保留关键字,但枚举值会转换为字符串 "global" hybrid = "hybrid" + mix = "mix" # Ollama API compatible models class OllamaMessage(BaseModel): @@ -684,7 +685,8 @@ def create_app(args): "/local ": SearchMode.local, "/global ": SearchMode.global_, # global_ is used because 'global' is a Python keyword "/naive ": SearchMode.naive, - "/hybrid ": SearchMode.hybrid + "/hybrid ": SearchMode.hybrid, + "/mix ": SearchMode.mix } for prefix, mode in mode_map.items(): diff --git a/test_lightrag_ollama_chat.py b/test_lightrag_ollama_chat.py index 5f8a03da..aab059aa 100644 --- a/test_lightrag_ollama_chat.py +++ b/test_lightrag_ollama_chat.py @@ -337,7 +337,7 @@ def test_query_modes(): 每个模式都会返回相同格式的响应,但检索策略不同。 """ url = get_base_url() - modes = ["local", "global", "naive", "hybrid"] # 支持的查询模式 + modes = ["local", "global", "naive", "hybrid", "mix"] # 支持的查询模式 for mode in modes: if OutputControl.is_verbose():