Improve OpenAI LLM logging with more detailed debug information

This commit is contained in:
yangdx
2025-03-28 21:33:59 +08:00
parent 5cd44047a2
commit 491c78dac1

View File

@@ -90,11 +90,13 @@ async def openai_complete_if_cache(
messages.extend(history_messages) messages.extend(history_messages)
messages.append({"role": "user", "content": prompt}) messages.append({"role": "user", "content": prompt})
logger.debug("===== Sending Query to LLM =====") logger.debug("===== Entering func of LLM =====")
logger.debug(f"Model: {model} Base URL: {base_url}") logger.debug(f"Model: {model} Base URL: {base_url}")
logger.debug(f"Additional kwargs: {kwargs}") logger.debug(f"Additional kwargs: {kwargs}")
verbose_debug(f"Query: {prompt}") logger.debug(f"Num of history messages: {len(history_messages)}")
verbose_debug(f"System prompt: {system_prompt}") verbose_debug(f"System prompt: {system_prompt}")
verbose_debug(f"Query: {prompt}")
logger.debug("===== Sending Query to LLM =====")
try: try:
if "response_format" in kwargs: if "response_format" in kwargs:
@@ -164,6 +166,9 @@ async def openai_complete_if_cache(
} }
token_tracker.add_usage(token_counts) token_tracker.add_usage(token_counts)
logger.debug(f"Response content len: {len(content)}")
verbose_debug(f"Response: {response}")
return content return content