Add debug logging for client configs in OpenAI LLM function
This commit is contained in:
parent
c87eb2cfcf
commit
451f488f72
1 changed files with 1 additions and 0 deletions
|
|
@ -172,6 +172,7 @@ async def openai_complete_if_cache(
|
|||
|
||||
logger.debug("===== Entering func of LLM =====")
|
||||
logger.debug(f"Model: {model} Base URL: {base_url}")
|
||||
logger.debug(f"Client Configs: {client_configs}")
|
||||
logger.debug(f"Additional kwargs: {kwargs}")
|
||||
logger.debug(f"Num of history messages: {len(history_messages)}")
|
||||
verbose_debug(f"System prompt: {system_prompt}")
|
||||
|
|
|
|||
Loading…
Add table
Reference in a new issue