- Set default max_tokens to 16384 (16K) for OpenAIGenericClient to better support local models - Add documentation note clarifying OpenAIGenericClient should be used for Ollama and LM Studio - Previous default was 8192 (8K) |
||
|---|---|---|
| .. | ||
| __init__.py | ||
| anthropic_client.py | ||
| azure_openai_client.py | ||
| client.py | ||
| config.py | ||
| errors.py | ||
| gemini_client.py | ||
| groq_client.py | ||
| openai_base_client.py | ||
| openai_client.py | ||
| openai_generic_client.py | ||
| utils.py | ||