diff --git a/env.example b/env.example index b322baa7..e18ee2fb 100644 --- a/env.example +++ b/env.example @@ -174,9 +174,12 @@ LLM_BINDING_API_KEY=your_api_key # LLM_BINDING_API_KEY=your_api_key # LLM_BINDING=openai -### OpenAI Specific Parameters -### Set the max_output_tokens to mitigate endless output of some LLM (less than LLM_TIMEOUT * llm_output_tokens/second, i.e. 9000 = 180s * 50 tokens/s) +### OpenAI Compatible API Specific Parameters +### Set the max_tokens to mitigate endless output of some LLM (less than LLM_TIMEOUT * llm_output_tokens/second, i.e. 9000 = 180s * 50 tokens/s) +### Typically, max_tokens does not include prompt content, though some models, such as Gemini-2.5-Flash, are exceptions +#### OpenAI's new API utilizes max_completion_tokens instead of max_tokens # OPENAI_LLM_MAX_TOKENS=9000 +# OPENAI_LLM_MAX_COMPLETION_TOKENS=9000 ### OpenRouter Specific Parameters # OPENAI_LLM_EXTRA_BODY='{"reasoning": {"enabled": false}}'