From 73c97771282c2a16b125b64084708c500f93559c Mon Sep 17 00:00:00 2001 From: Igor Ilic Date: Mon, 8 Sep 2025 21:22:50 +0200 Subject: [PATCH 1/2] refactor: Set Cognee to use gpt4o-mini by default --- cognee/infrastructure/llm/config.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/cognee/infrastructure/llm/config.py b/cognee/infrastructure/llm/config.py index 7aa8f33f7..35c941e89 100644 --- a/cognee/infrastructure/llm/config.py +++ b/cognee/infrastructure/llm/config.py @@ -35,7 +35,7 @@ class LLMConfig(BaseSettings): structured_output_framework: str = "instructor" llm_provider: str = "openai" - llm_model: str = "gpt-5-mini" + llm_model: str = "openai/gpt-4o-mini" llm_endpoint: str = "" llm_api_key: Optional[str] = None llm_api_version: Optional[str] = None @@ -44,7 +44,7 @@ class LLMConfig(BaseSettings): llm_max_completion_tokens: int = 16384 baml_llm_provider: str = "openai" - baml_llm_model: str = "gpt-5-mini" + baml_llm_model: str = "openai/gpt-4o-mini" baml_llm_endpoint: str = "" baml_llm_api_key: Optional[str] = None baml_llm_temperature: float = 0.0 From 47b62d50e4e2432813d9372d577aed40d507c10c Mon Sep 17 00:00:00 2001 From: Igor Ilic Date: Mon, 8 Sep 2025 21:27:38 +0200 Subject: [PATCH 2/2] refactor: Use gpt-4o-mini by default --- .env.template | 4 ++-- cognee/infrastructure/llm/config.py | 2 +- 2 files changed, 3 insertions(+), 3 deletions(-) diff --git a/.env.template b/.env.template index ee62f1d3d..28980de74 100644 --- a/.env.template +++ b/.env.template @@ -16,7 +16,7 @@ STRUCTURED_OUTPUT_FRAMEWORK="instructor" LLM_API_KEY="your_api_key" -LLM_MODEL="openai/gpt-5-mini" +LLM_MODEL="openai/gpt-4o-mini" LLM_PROVIDER="openai" LLM_ENDPOINT="" LLM_API_VERSION="" @@ -33,7 +33,7 @@ EMBEDDING_MAX_TOKENS=8191 # If using BAML structured output these env variables will be used BAML_LLM_PROVIDER=openai -BAML_LLM_MODEL="gpt-5-mini" +BAML_LLM_MODEL="gpt-4o-mini" BAML_LLM_ENDPOINT="" BAML_LLM_API_KEY="your_api_key" BAML_LLM_API_VERSION="" diff --git a/cognee/infrastructure/llm/config.py b/cognee/infrastructure/llm/config.py index 35c941e89..b2fd460f2 100644 --- a/cognee/infrastructure/llm/config.py +++ b/cognee/infrastructure/llm/config.py @@ -44,7 +44,7 @@ class LLMConfig(BaseSettings): llm_max_completion_tokens: int = 16384 baml_llm_provider: str = "openai" - baml_llm_model: str = "openai/gpt-4o-mini" + baml_llm_model: str = "gpt-4o-mini" baml_llm_endpoint: str = "" baml_llm_api_key: Optional[str] = None baml_llm_temperature: float = 0.0