Improve OpenAI LLM logging with more detailed debug information
This commit is contained in:
parent
5cd44047a2
commit
491c78dac1
1 changed files with 7 additions and 2 deletions
|
|
@ -90,11 +90,13 @@ async def openai_complete_if_cache(
|
||||||
messages.extend(history_messages)
|
messages.extend(history_messages)
|
||||||
messages.append({"role": "user", "content": prompt})
|
messages.append({"role": "user", "content": prompt})
|
||||||
|
|
||||||
logger.debug("===== Sending Query to LLM =====")
|
logger.debug("===== Entering func of LLM =====")
|
||||||
logger.debug(f"Model: {model} Base URL: {base_url}")
|
logger.debug(f"Model: {model} Base URL: {base_url}")
|
||||||
logger.debug(f"Additional kwargs: {kwargs}")
|
logger.debug(f"Additional kwargs: {kwargs}")
|
||||||
verbose_debug(f"Query: {prompt}")
|
logger.debug(f"Num of history messages: {len(history_messages)}")
|
||||||
verbose_debug(f"System prompt: {system_prompt}")
|
verbose_debug(f"System prompt: {system_prompt}")
|
||||||
|
verbose_debug(f"Query: {prompt}")
|
||||||
|
logger.debug("===== Sending Query to LLM =====")
|
||||||
|
|
||||||
try:
|
try:
|
||||||
if "response_format" in kwargs:
|
if "response_format" in kwargs:
|
||||||
|
|
@ -164,6 +166,9 @@ async def openai_complete_if_cache(
|
||||||
}
|
}
|
||||||
token_tracker.add_usage(token_counts)
|
token_tracker.add_usage(token_counts)
|
||||||
|
|
||||||
|
logger.debug(f"Response content len: {len(content)}")
|
||||||
|
verbose_debug(f"Response: {response}")
|
||||||
|
|
||||||
return content
|
return content
|
||||||
|
|
||||||
|
|
||||||
|
|
|
||||||
Loading…
Add table
Reference in a new issue