Compare commits

...
Sign in to create a new pull request.

1 commit

Author SHA1 Message Date
claude[bot]
36bca7ba10 fix: update default Gemini model from deprecated gemini-2.5-flash-lite-preview-06-17 to gemini-2.5-flash-lite
- Updated DEFAULT_SMALL_MODEL in gemini_client.py
- Updated DEFAULT_MODEL in gemini_reranker_client.py
- Removed deprecated model from GEMINI_MODEL_MAX_TOKENS mapping
- Updated README.md examples and documentation
- Cleaned up test cases in test_gemini_client.py

Fixes #1075

Co-authored-by: Daniel Chalef <danielchalef@users.noreply.github.com>
2025-11-20 23:33:59 +00:00
4 changed files with 4 additions and 6 deletions

View file

@ -479,7 +479,7 @@ graphiti = Graphiti(
cross_encoder=GeminiRerankerClient( cross_encoder=GeminiRerankerClient(
config=LLMConfig( config=LLMConfig(
api_key=api_key, api_key=api_key,
model="gemini-2.5-flash-lite-preview-06-17" model="gemini-2.5-flash-lite"
) )
) )
) )
@ -487,7 +487,7 @@ graphiti = Graphiti(
# Now you can use Graphiti with Google Gemini for all components # Now you can use Graphiti with Google Gemini for all components
``` ```
The Gemini reranker uses the `gemini-2.5-flash-lite-preview-06-17` model by default, which is optimized for The Gemini reranker uses the `gemini-2.5-flash-lite` model by default, which is optimized for
cost-effective and low-latency classification tasks. It uses the same boolean classification approach as the OpenAI cost-effective and low-latency classification tasks. It uses the same boolean classification approach as the OpenAI
reranker, leveraging Gemini's log probabilities feature to rank passage relevance. reranker, leveraging Gemini's log probabilities feature to rank passage relevance.

View file

@ -37,7 +37,7 @@ else:
logger = logging.getLogger(__name__) logger = logging.getLogger(__name__)
DEFAULT_MODEL = 'gemini-2.5-flash-lite-preview-06-17' DEFAULT_MODEL = 'gemini-2.5-flash-lite'
class GeminiRerankerClient(CrossEncoderClient): class GeminiRerankerClient(CrossEncoderClient):

View file

@ -45,7 +45,7 @@ else:
logger = logging.getLogger(__name__) logger = logging.getLogger(__name__)
DEFAULT_MODEL = 'gemini-2.5-flash' DEFAULT_MODEL = 'gemini-2.5-flash'
DEFAULT_SMALL_MODEL = 'gemini-2.5-flash-lite-preview-06-17' DEFAULT_SMALL_MODEL = 'gemini-2.5-flash-lite'
# Maximum output tokens for different Gemini models # Maximum output tokens for different Gemini models
GEMINI_MODEL_MAX_TOKENS = { GEMINI_MODEL_MAX_TOKENS = {
@ -53,7 +53,6 @@ GEMINI_MODEL_MAX_TOKENS = {
'gemini-2.5-pro': 65536, 'gemini-2.5-pro': 65536,
'gemini-2.5-flash': 65536, 'gemini-2.5-flash': 65536,
'gemini-2.5-flash-lite': 64000, 'gemini-2.5-flash-lite': 64000,
'models/gemini-2.5-flash-lite-preview-06-17': 64000,
# Gemini 2.0 models # Gemini 2.0 models
'gemini-2.0-flash': 8192, 'gemini-2.0-flash': 8192,
'gemini-2.0-flash-lite': 8192, 'gemini-2.0-flash-lite': 8192,

View file

@ -455,7 +455,6 @@ class TestGeminiClientGenerateResponse:
('gemini-2.5-flash', 65536), ('gemini-2.5-flash', 65536),
('gemini-2.5-pro', 65536), ('gemini-2.5-pro', 65536),
('gemini-2.5-flash-lite', 64000), ('gemini-2.5-flash-lite', 64000),
('models/gemini-2.5-flash-lite-preview-06-17', 64000),
('gemini-2.0-flash', 8192), ('gemini-2.0-flash', 8192),
('gemini-1.5-pro', 8192), ('gemini-1.5-pro', 8192),
('gemini-1.5-flash', 8192), ('gemini-1.5-flash', 8192),