From ca0092d51094ccd1ae2fc27d0bf3faf71119e91d Mon Sep 17 00:00:00 2001 From: Daniel Chalef <131175+danielchalef@users.noreply.github.com> Date: Thu, 30 Oct 2025 10:41:47 -0700 Subject: [PATCH] conductor-checkpoint-msg_018FX6Mibr66cKLnpL84f2Js --- graphiti_core/llm_client/config.py | 2 +- mcp_server/src/config/schema.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/graphiti_core/llm_client/config.py b/graphiti_core/llm_client/config.py index dec7a293..17a6d5a6 100644 --- a/graphiti_core/llm_client/config.py +++ b/graphiti_core/llm_client/config.py @@ -39,7 +39,7 @@ class LLMConfig: api_key: str | None = None, model: str | None = None, base_url: str | None = None, - temperature: float = DEFAULT_TEMPERATURE, + temperature: float | None = None, max_tokens: int = DEFAULT_MAX_TOKENS, small_model: str | None = None, ): diff --git a/mcp_server/src/config/schema.py b/mcp_server/src/config/schema.py index e3579041..5848bfff 100644 --- a/mcp_server/src/config/schema.py +++ b/mcp_server/src/config/schema.py @@ -151,7 +151,7 @@ class LLMConfig(BaseModel): provider: str = Field(default='openai', description='LLM provider') model: str = Field(default='gpt-4.1', description='Model name') - temperature: float = Field(default=0.0, description='Temperature') + temperature: float | None = Field(default=None, description='Temperature (optional, defaults to None for reasoning models)') max_tokens: int = Field(default=4096, description='Max tokens') providers: LLMProvidersConfig = Field(default_factory=LLMProvidersConfig)