graphiti/graphiti_core/llm_client/config.py
facronactz a7f2c92bb4 feat: Add support for Chutes API integration
- Introduced CHUTES_API_TOKEN in .env.example files for configuration.
- Updated README.md in quickstart example to include Chutes API setup.
- Created quickstart_chutes.py example for demonstrating Chutes integration.
- Added ChutesEmbedder and ChutesClient classes for embedding and LLM functionalities.
- Updated LLM and embedder provider configurations to support Chutes.
- Enhanced factories to include Chutes client and embedder.
- Modified schema.py to define Chutes provider configuration.
- Updated mcp_server configuration files to integrate Chutes API.
- Added necessary dependencies in pyproject.toml for Chutes support.
2025-11-15 16:12:09 +07:00

77 lines
2.8 KiB
Python

"""
Copyright 2024, Zep Software, Inc.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
"""
from enum import Enum
DEFAULT_MAX_TOKENS = 8192
DEFAULT_TEMPERATURE = 1
class ModelSize(Enum):
small = 'small'
medium = 'medium'
class LLMProvider(str, Enum):
OPENAI = "openai"
CHUTES = "chutes"
class LLMConfig:
"""
Configuration class for the Language Learning Model (LLM).
This class encapsulates the necessary parameters to interact with an LLM API,
such as OpenAI's GPT models. It stores the API key, model name, and base URL
for making requests to the LLM service.
"""
def __init__(
self,
api_key: str | None = None,
model: str | None = None,
base_url: str | None = None,
temperature: float = DEFAULT_TEMPERATURE,
max_tokens: int = DEFAULT_MAX_TOKENS,
small_model: str | None = None,
provider: LLMProvider = LLMProvider.OPENAI,
):
"""
Initialize the LLMConfig with the provided parameters.
Args:
api_key (str): The authentication key for accessing the LLM API.
This is required for making authorized requests.
model (str, optional): The specific LLM model to use for generating responses.
Defaults to "gpt-4.1-mini".
base_url (str, optional): The base URL of the LLM API service.
Defaults to "https://api.openai.com", which is OpenAI's standard API endpoint.
This can be changed if using a different provider or a custom endpoint.
small_model (str, optional): The specific LLM model to use for generating responses of simpler prompts.
Defaults to "gpt-4.1-nano".
provider (LLMProvider, optional): The LLM provider to use.
Defaults to LLMProvider.OPENAI.
"""
self.base_url = base_url
self.api_key = api_key
self.model = model
self.small_model = small_model
self.temperature = temperature
self.max_tokens = max_tokens
self.provider = provider