fix: Resolve issue with BAML not working without LLM_API_KEY set fro LiteLLM

This commit is contained in:
Igor Ilic 2025-09-09 18:11:34 +02:00
parent 4258f2e92b
commit 4bae611721
2 changed files with 6 additions and 6 deletions

View file

@ -32,7 +32,7 @@ class LLMProvider(Enum):
GEMINI = "gemini"
def get_llm_client():
def get_llm_client(raise_api_key_error: bool = True):
"""
Get the LLM client based on the configuration using Enums.
@ -65,7 +65,7 @@ def get_llm_client():
)
if provider == LLMProvider.OPENAI:
if llm_config.llm_api_key is None:
if llm_config.llm_api_key is None and raise_api_key_error:
raise LLMAPIKeyNotSetError()
from cognee.infrastructure.llm.structured_output_framework.litellm_instructor.llm.openai.adapter import (
@ -86,7 +86,7 @@ def get_llm_client():
)
elif provider == LLMProvider.OLLAMA:
if llm_config.llm_api_key is None:
if llm_config.llm_api_key is None and raise_api_key_error:
raise LLMAPIKeyNotSetError()
from cognee.infrastructure.llm.structured_output_framework.litellm_instructor.llm.generic_llm_api.adapter import (
@ -111,7 +111,7 @@ def get_llm_client():
)
elif provider == LLMProvider.CUSTOM:
if llm_config.llm_api_key is None:
if llm_config.llm_api_key is None and raise_api_key_error:
raise LLMAPIKeyNotSetError()
from cognee.infrastructure.llm.structured_output_framework.litellm_instructor.llm.generic_llm_api.adapter import (
@ -130,7 +130,7 @@ def get_llm_client():
)
elif provider == LLMProvider.GEMINI:
if llm_config.llm_api_key is None:
if llm_config.llm_api_key is None and raise_api_key_error:
raise LLMAPIKeyNotSetError()
from cognee.infrastructure.llm.structured_output_framework.litellm_instructor.llm.gemini.adapter import (

View file

@ -29,7 +29,7 @@ def get_max_chunk_tokens():
# Calculate max chunk size based on the following formula
embedding_engine = get_vector_engine().embedding_engine
llm_client = get_llm_client()
llm_client = get_llm_client(raise_api_key_error=False)
# We need to make sure chunk size won't take more than half of LLM max context token size
# but it also can't be bigger than the embedding engine max token size