* feat: enhance GeminiClient with max tokens management - Introduced a mapping for maximum output tokens for various Gemini models. - Added methods to resolve max tokens based on precedence rules, allowing for more flexible token management. - Updated tests to verify max tokens behavior, ensuring explicit parameters take precedence and fallback mechanisms work correctly. This change improves the handling of token limits for different models, enhancing the client’s configurability and usability. * refactor: streamline max tokens retrieval in GeminiClient - Removed the fallback to DEFAULT_MAX_TOKENS in favor of directly using model-specific maximum tokens. - Simplified the logic for determining max tokens, enhancing code clarity and maintainability. This change improves the efficiency of token management within the GeminiClient. |
||
|---|---|---|
| .. | ||
| cross_encoder | ||
| driver | ||
| embedder | ||
| llm_client | ||
| models | ||
| prompts | ||
| search | ||
| telemetry | ||
| utils | ||
| __init__.py | ||
| edges.py | ||
| errors.py | ||
| graph_queries.py | ||
| graphiti.py | ||
| graphiti_types.py | ||
| helpers.py | ||
| nodes.py | ||
| py.typed | ||