Add support for Ollama streaming output and integrate Open-WebUI as the chat UI demo
This commit is contained in:
@@ -19,6 +19,7 @@ class QueryParam:
|
||||
only_need_context: bool = False
|
||||
only_need_prompt: bool = False
|
||||
response_type: str = "Multiple Paragraphs"
|
||||
stream: bool = False
|
||||
# Number of top-k items to retrieve; corresponds to entities in "local" mode and relationships in "global" mode.
|
||||
top_k: int = 60
|
||||
# Number of document chunks to retrieve.
|
||||
|
Reference in New Issue
Block a user