From fada4ffd340e2610de60981497d2d9b56888b024 Mon Sep 17 00:00:00 2001 From: "claude[bot]" <41898282+claude[bot]@users.noreply.github.com> Date: Thu, 20 Nov 2025 23:36:50 +0000 Subject: [PATCH] fix: replace deprecated gemini-2.5-flash-lite-preview-06-17 with gemini-2.5-flash-lite MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Updated all references to the deprecated Gemini model in: - graphiti_core/llm_client/gemini_client.py - graphiti_core/cross_encoder/gemini_reranker_client.py - tests/llm_client/test_gemini_client.py - README.md This resolves 404 errors when using Gemini clients. Fixes #1075 🤖 Generated with [Claude Code](https://claude.ai/code) Co-authored-by: Daniel Chalef --- README.md | 4 ++-- graphiti_core/cross_encoder/gemini_reranker_client.py | 2 +- graphiti_core/llm_client/gemini_client.py | 3 +-- tests/llm_client/test_gemini_client.py | 1 - 4 files changed, 4 insertions(+), 6 deletions(-) diff --git a/README.md b/README.md index 1efd37df..57259675 100644 --- a/README.md +++ b/README.md @@ -479,7 +479,7 @@ graphiti = Graphiti( cross_encoder=GeminiRerankerClient( config=LLMConfig( api_key=api_key, - model="gemini-2.5-flash-lite-preview-06-17" + model="gemini-2.5-flash-lite" ) ) ) @@ -487,7 +487,7 @@ graphiti = Graphiti( # Now you can use Graphiti with Google Gemini for all components ``` -The Gemini reranker uses the `gemini-2.5-flash-lite-preview-06-17` model by default, which is optimized for +The Gemini reranker uses the `gemini-2.5-flash-lite` model by default, which is optimized for cost-effective and low-latency classification tasks. It uses the same boolean classification approach as the OpenAI reranker, leveraging Gemini's log probabilities feature to rank passage relevance. diff --git a/graphiti_core/cross_encoder/gemini_reranker_client.py b/graphiti_core/cross_encoder/gemini_reranker_client.py index 01f13b9d..b51892b4 100644 --- a/graphiti_core/cross_encoder/gemini_reranker_client.py +++ b/graphiti_core/cross_encoder/gemini_reranker_client.py @@ -37,7 +37,7 @@ else: logger = logging.getLogger(__name__) -DEFAULT_MODEL = 'gemini-2.5-flash-lite-preview-06-17' +DEFAULT_MODEL = 'gemini-2.5-flash-lite' class GeminiRerankerClient(CrossEncoderClient): diff --git a/graphiti_core/llm_client/gemini_client.py b/graphiti_core/llm_client/gemini_client.py index 1973a98f..46b4c8d1 100644 --- a/graphiti_core/llm_client/gemini_client.py +++ b/graphiti_core/llm_client/gemini_client.py @@ -45,7 +45,7 @@ else: logger = logging.getLogger(__name__) DEFAULT_MODEL = 'gemini-2.5-flash' -DEFAULT_SMALL_MODEL = 'gemini-2.5-flash-lite-preview-06-17' +DEFAULT_SMALL_MODEL = 'gemini-2.5-flash-lite' # Maximum output tokens for different Gemini models GEMINI_MODEL_MAX_TOKENS = { @@ -53,7 +53,6 @@ GEMINI_MODEL_MAX_TOKENS = { 'gemini-2.5-pro': 65536, 'gemini-2.5-flash': 65536, 'gemini-2.5-flash-lite': 64000, - 'models/gemini-2.5-flash-lite-preview-06-17': 64000, # Gemini 2.0 models 'gemini-2.0-flash': 8192, 'gemini-2.0-flash-lite': 8192, diff --git a/tests/llm_client/test_gemini_client.py b/tests/llm_client/test_gemini_client.py index 263f93c6..e0178253 100644 --- a/tests/llm_client/test_gemini_client.py +++ b/tests/llm_client/test_gemini_client.py @@ -455,7 +455,6 @@ class TestGeminiClientGenerateResponse: ('gemini-2.5-flash', 65536), ('gemini-2.5-pro', 65536), ('gemini-2.5-flash-lite', 64000), - ('models/gemini-2.5-flash-lite-preview-06-17', 64000), ('gemini-2.0-flash', 8192), ('gemini-1.5-pro', 8192), ('gemini-1.5-flash', 8192),