fix: Move to gpt5 (#1262)
<!-- .github/pull_request_template.md --> ## Description <!-- Provide a clear description of the changes in this PR --> ## DCO Affirmation I affirm that all code in every commit of this pull request conforms to the terms of the Topoteretes Developer Certificate of Origin.
This commit is contained in:
commit
abbce6e8c3
41 changed files with 541 additions and 1215 deletions
|
|
@ -6,7 +6,7 @@
|
||||||
# Default graph database : Kuzu
|
# Default graph database : Kuzu
|
||||||
#
|
#
|
||||||
# These default databases are all file-based, so no extra setup is needed
|
# These default databases are all file-based, so no extra setup is needed
|
||||||
# for local use.
|
# for local use. The data by default will be stored in your .venv
|
||||||
###############################################################################
|
###############################################################################
|
||||||
|
|
||||||
################################################################################
|
################################################################################
|
||||||
|
|
@ -16,7 +16,7 @@
|
||||||
STRUCTURED_OUTPUT_FRAMEWORK="instructor"
|
STRUCTURED_OUTPUT_FRAMEWORK="instructor"
|
||||||
|
|
||||||
LLM_API_KEY="your_api_key"
|
LLM_API_KEY="your_api_key"
|
||||||
LLM_MODEL="openai/gpt-4o-mini"
|
LLM_MODEL="openai/gpt-5-mini"
|
||||||
LLM_PROVIDER="openai"
|
LLM_PROVIDER="openai"
|
||||||
LLM_ENDPOINT=""
|
LLM_ENDPOINT=""
|
||||||
LLM_API_VERSION=""
|
LLM_API_VERSION=""
|
||||||
|
|
@ -33,11 +33,20 @@ EMBEDDING_MAX_TOKENS=8191
|
||||||
|
|
||||||
# If using BAML structured output these env variables will be used
|
# If using BAML structured output these env variables will be used
|
||||||
BAML_LLM_PROVIDER=openai
|
BAML_LLM_PROVIDER=openai
|
||||||
BAML_LLM_MODEL="gpt-4o-mini"
|
BAML_LLM_MODEL="gpt-5-mini"
|
||||||
BAML_LLM_ENDPOINT=""
|
BAML_LLM_ENDPOINT=""
|
||||||
BAML_LLM_API_KEY="your_api_key"
|
BAML_LLM_API_KEY="your_api_key"
|
||||||
BAML_LLM_API_VERSION=""
|
BAML_LLM_API_VERSION=""
|
||||||
|
|
||||||
|
|
||||||
|
################################################################################
|
||||||
|
# 📂 ROOT DIRECTORY FOR DATABASES
|
||||||
|
################################################################################
|
||||||
|
# Set up the Cognee system directory. Cognee will store system files and databases here.
|
||||||
|
# Useful for setting root directory inside docker and also to avoid storing the databases in .venv
|
||||||
|
# DATA_ROOT_DIRECTORY='/Users/<user>/Desktop/cognee/.cognee_data/'
|
||||||
|
# SYSTEM_ROOT_DIRECTORY='/Users/<user>/Desktop/cognee/.cognee_system/'
|
||||||
|
|
||||||
################################################################################
|
################################################################################
|
||||||
# 🗄️ Relational database settings
|
# 🗄️ Relational database settings
|
||||||
################################################################################
|
################################################################################
|
||||||
|
|
@ -85,12 +94,7 @@ VECTOR_DB_PROVIDER="lancedb"
|
||||||
VECTOR_DB_URL=
|
VECTOR_DB_URL=
|
||||||
VECTOR_DB_KEY=
|
VECTOR_DB_KEY=
|
||||||
|
|
||||||
################################################################################
|
|
||||||
# 📂 ROOT DIRECTORY IF USING COGNEE LIB INSIDE A DOCKER
|
|
||||||
################################################################################
|
|
||||||
# Set up the Cognee system directory. Cognee will store system files and databases here.
|
|
||||||
# DATA_ROOT_DIRECTORY='/Users/<user>/Desktop/cognee/.cognee_data/'
|
|
||||||
# SYSTEM_ROOT_DIRECTORY='/Users/<user>/Desktop/cognee/.cognee_system/'
|
|
||||||
|
|
||||||
################################################################################
|
################################################################################
|
||||||
# 🔄 MIGRATION (RELATIONAL → GRAPH) SETTINGS
|
# 🔄 MIGRATION (RELATIONAL → GRAPH) SETTINGS
|
||||||
|
|
@ -157,8 +161,8 @@ LITELLM_LOG="ERROR"
|
||||||
# Uncomment + fill values to switch.
|
# Uncomment + fill values to switch.
|
||||||
|
|
||||||
########## Azure OpenAI #######################################################
|
########## Azure OpenAI #######################################################
|
||||||
#LLM_MODEL="azure/gpt-4o-mini"
|
#LLM_MODEL="azure/gpt-5-mini"
|
||||||
#LLM_ENDPOINT="https://DNS.azure.com/openai/deployments/gpt-4o-mini"
|
#LLM_ENDPOINT="https://DNS.azure.com/openai/deployments/gpt-5-mini"
|
||||||
#LLM_API_KEY="<<TALK TO YOUR AZURE GUY"
|
#LLM_API_KEY="<<TALK TO YOUR AZURE GUY"
|
||||||
#LLM_API_VERSION="2024-12-01-preview"
|
#LLM_API_VERSION="2024-12-01-preview"
|
||||||
|
|
||||||
|
|
|
||||||
6
.github/workflows/weighted_edges_tests.yml
vendored
6
.github/workflows/weighted_edges_tests.yml
vendored
|
|
@ -31,7 +31,7 @@ jobs:
|
||||||
python-version: ['3.11', '3.12']
|
python-version: ['3.11', '3.12']
|
||||||
env:
|
env:
|
||||||
LLM_PROVIDER: openai
|
LLM_PROVIDER: openai
|
||||||
LLM_MODEL: gpt-4o-mini
|
LLM_MODEL: gpt-5-mini
|
||||||
LLM_API_KEY: ${{ secrets.LLM_API_KEY }}
|
LLM_API_KEY: ${{ secrets.LLM_API_KEY }}
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
|
|
@ -66,7 +66,7 @@ jobs:
|
||||||
graph_db_provider: "neo4j"
|
graph_db_provider: "neo4j"
|
||||||
env:
|
env:
|
||||||
LLM_PROVIDER: openai
|
LLM_PROVIDER: openai
|
||||||
LLM_MODEL: gpt-4o-mini
|
LLM_MODEL: gpt-5-mini
|
||||||
LLM_ENDPOINT: https://api.openai.com/v1/
|
LLM_ENDPOINT: https://api.openai.com/v1/
|
||||||
LLM_API_KEY: ${{ secrets.LLM_API_KEY }}
|
LLM_API_KEY: ${{ secrets.LLM_API_KEY }}
|
||||||
LLM_API_VERSION: "2024-02-01"
|
LLM_API_VERSION: "2024-02-01"
|
||||||
|
|
@ -100,7 +100,7 @@ jobs:
|
||||||
runs-on: ubuntu-22.04
|
runs-on: ubuntu-22.04
|
||||||
env:
|
env:
|
||||||
LLM_PROVIDER: openai
|
LLM_PROVIDER: openai
|
||||||
LLM_MODEL: gpt-4o-mini
|
LLM_MODEL: gpt-5-mini
|
||||||
LLM_ENDPOINT: https://api.openai.com/v1/
|
LLM_ENDPOINT: https://api.openai.com/v1/
|
||||||
LLM_API_KEY: ${{ secrets.LLM_API_KEY }}
|
LLM_API_KEY: ${{ secrets.LLM_API_KEY }}
|
||||||
LLM_API_VERSION: "2024-02-01"
|
LLM_API_VERSION: "2024-02-01"
|
||||||
|
|
|
||||||
|
|
@ -91,7 +91,8 @@ Your contributions are at the core of making this a true open source project. An
|
||||||
|
|
||||||
## 📦 Installation
|
## 📦 Installation
|
||||||
|
|
||||||
You can install Cognee using either **uv**, **pip**, **poetry** or any other python package manager.
|
You can install Cognee using either **pip**, **poetry**, **uv** or any other python package manager.
|
||||||
|
|
||||||
Cognee supports Python 3.10 to 3.13
|
Cognee supports Python 3.10 to 3.13
|
||||||
|
|
||||||
### With pip
|
### With pip
|
||||||
|
|
|
||||||
Binary file not shown.
|
Before Width: | Height: | Size: 2.2 MiB After Width: | Height: | Size: 63 KiB |
|
|
@ -34,8 +34,8 @@
|
||||||
// };
|
// };
|
||||||
|
|
||||||
// const defaultModel = {
|
// const defaultModel = {
|
||||||
// label: 'gpt-4o-mini',
|
// label: 'gpt-5-mini',
|
||||||
// value: 'gpt-4o-mini',
|
// value: 'gpt-5-mini',
|
||||||
// };
|
// };
|
||||||
|
|
||||||
// export default function Settings({ onDone = () => {}, submitButtonText = 'Save' }) {
|
// export default function Settings({ onDone = () => {}, submitButtonText = 'Save' }) {
|
||||||
|
|
|
||||||
|
|
@ -8,7 +8,7 @@ requires-python = ">=3.10"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
# For local cognee repo usage remove comment bellow and add absolute path to cognee. Then run `uv sync --reinstall` in the mcp folder on local cognee changes.
|
# For local cognee repo usage remove comment bellow and add absolute path to cognee. Then run `uv sync --reinstall` in the mcp folder on local cognee changes.
|
||||||
# "cognee[postgres,codegraph,gemini,huggingface,docs,neo4j] @ file:/Users/vasilije/Projects/tiktok/cognee",
|
# "cognee[postgres,codegraph,gemini,huggingface,docs,neo4j] @ file:/Users/vasilije/Projects/tiktok/cognee",
|
||||||
"cognee[postgres,codegraph,gemini,huggingface,docs,neo4j]==0.2.1",
|
"cognee[postgres,codegraph,gemini,huggingface,docs,neo4j]==0.2.2",
|
||||||
"fastmcp>=2.10.0,<3.0.0",
|
"fastmcp>=2.10.0,<3.0.0",
|
||||||
"mcp>=1.12.0,<2.0.0",
|
"mcp>=1.12.0,<2.0.0",
|
||||||
"uv>=0.6.3,<1.0.0",
|
"uv>=0.6.3,<1.0.0",
|
||||||
|
|
|
||||||
|
|
@ -1,6 +1,6 @@
|
||||||
# In case you choose to use OpenAI provider, just adjust the model and api_key.
|
# In case you choose to use OpenAI provider, just adjust the model and api_key.
|
||||||
LLM_API_KEY=""
|
LLM_API_KEY=""
|
||||||
LLM_MODEL="openai/gpt-4o-mini"
|
LLM_MODEL="openai/gpt-5-mini"
|
||||||
LLM_PROVIDER="openai"
|
LLM_PROVIDER="openai"
|
||||||
# Not needed if you use OpenAI
|
# Not needed if you use OpenAI
|
||||||
LLM_ENDPOINT=""
|
LLM_ENDPOINT=""
|
||||||
|
|
|
||||||
|
|
@ -128,7 +128,7 @@ async def add(
|
||||||
|
|
||||||
Optional:
|
Optional:
|
||||||
- LLM_PROVIDER: "openai" (default), "anthropic", "gemini", "ollama"
|
- LLM_PROVIDER: "openai" (default), "anthropic", "gemini", "ollama"
|
||||||
- LLM_MODEL: Model name (default: "gpt-4o-mini")
|
- LLM_MODEL: Model name (default: "gpt-5-mini")
|
||||||
- DEFAULT_USER_EMAIL: Custom default user email
|
- DEFAULT_USER_EMAIL: Custom default user email
|
||||||
- DEFAULT_USER_PASSWORD: Custom default user password
|
- DEFAULT_USER_PASSWORD: Custom default user password
|
||||||
- VECTOR_DB_PROVIDER: "lancedb" (default), "chromadb", "pgvector"
|
- VECTOR_DB_PROVIDER: "lancedb" (default), "chromadb", "pgvector"
|
||||||
|
|
|
||||||
|
|
@ -91,7 +91,7 @@ async def cognify(
|
||||||
- LangchainChunker: Recursive character splitting with overlap
|
- LangchainChunker: Recursive character splitting with overlap
|
||||||
Determines how documents are segmented for processing.
|
Determines how documents are segmented for processing.
|
||||||
chunk_size: Maximum tokens per chunk. Auto-calculated based on LLM if None.
|
chunk_size: Maximum tokens per chunk. Auto-calculated based on LLM if None.
|
||||||
Formula: min(embedding_max_tokens, llm_max_tokens // 2)
|
Formula: min(embedding_max_completion_tokens, llm_max_completion_tokens // 2)
|
||||||
Default limits: ~512-8192 tokens depending on models.
|
Default limits: ~512-8192 tokens depending on models.
|
||||||
Smaller chunks = more granular but potentially fragmented knowledge.
|
Smaller chunks = more granular but potentially fragmented knowledge.
|
||||||
ontology_file_path: Path to RDF/OWL ontology file for domain-specific entity types.
|
ontology_file_path: Path to RDF/OWL ontology file for domain-specific entity types.
|
||||||
|
|
|
||||||
|
|
@ -70,7 +70,7 @@ class ResponseRequest(InDTO):
|
||||||
tool_choice: Optional[Union[str, Dict[str, Any]]] = "auto"
|
tool_choice: Optional[Union[str, Dict[str, Any]]] = "auto"
|
||||||
user: Optional[str] = None
|
user: Optional[str] = None
|
||||||
temperature: Optional[float] = 1.0
|
temperature: Optional[float] = 1.0
|
||||||
max_tokens: Optional[int] = None
|
max_completion_tokens: Optional[int] = None
|
||||||
|
|
||||||
|
|
||||||
class ToolCallOutput(BaseModel):
|
class ToolCallOutput(BaseModel):
|
||||||
|
|
|
||||||
|
|
@ -25,7 +25,7 @@ class EvalConfig(BaseSettings):
|
||||||
"EM",
|
"EM",
|
||||||
"f1",
|
"f1",
|
||||||
] # Use only 'correctness' for DirectLLM
|
] # Use only 'correctness' for DirectLLM
|
||||||
deepeval_model: str = "gpt-4o-mini"
|
deepeval_model: str = "gpt-5-mini"
|
||||||
|
|
||||||
# Metrics params
|
# Metrics params
|
||||||
calculate_metrics: bool = True
|
calculate_metrics: bool = True
|
||||||
|
|
|
||||||
|
|
@ -41,11 +41,11 @@ class FastembedEmbeddingEngine(EmbeddingEngine):
|
||||||
self,
|
self,
|
||||||
model: Optional[str] = "openai/text-embedding-3-large",
|
model: Optional[str] = "openai/text-embedding-3-large",
|
||||||
dimensions: Optional[int] = 3072,
|
dimensions: Optional[int] = 3072,
|
||||||
max_tokens: int = 512,
|
max_completion_tokens: int = 512,
|
||||||
):
|
):
|
||||||
self.model = model
|
self.model = model
|
||||||
self.dimensions = dimensions
|
self.dimensions = dimensions
|
||||||
self.max_tokens = max_tokens
|
self.max_completion_tokens = max_completion_tokens
|
||||||
self.tokenizer = self.get_tokenizer()
|
self.tokenizer = self.get_tokenizer()
|
||||||
# self.retry_count = 0
|
# self.retry_count = 0
|
||||||
self.embedding_model = TextEmbedding(model_name=model)
|
self.embedding_model = TextEmbedding(model_name=model)
|
||||||
|
|
@ -112,7 +112,9 @@ class FastembedEmbeddingEngine(EmbeddingEngine):
|
||||||
"""
|
"""
|
||||||
logger.debug("Loading tokenizer for FastembedEmbeddingEngine...")
|
logger.debug("Loading tokenizer for FastembedEmbeddingEngine...")
|
||||||
|
|
||||||
tokenizer = TikTokenTokenizer(model="gpt-4o", max_tokens=self.max_tokens)
|
tokenizer = TikTokenTokenizer(
|
||||||
|
model="gpt-4o", max_completion_tokens=self.max_completion_tokens
|
||||||
|
)
|
||||||
|
|
||||||
logger.debug("Tokenizer loaded for for FastembedEmbeddingEngine")
|
logger.debug("Tokenizer loaded for for FastembedEmbeddingEngine")
|
||||||
return tokenizer
|
return tokenizer
|
||||||
|
|
|
||||||
|
|
@ -57,7 +57,7 @@ class LiteLLMEmbeddingEngine(EmbeddingEngine):
|
||||||
api_key: str = None,
|
api_key: str = None,
|
||||||
endpoint: str = None,
|
endpoint: str = None,
|
||||||
api_version: str = None,
|
api_version: str = None,
|
||||||
max_tokens: int = 512,
|
max_completion_tokens: int = 512,
|
||||||
):
|
):
|
||||||
self.api_key = api_key
|
self.api_key = api_key
|
||||||
self.endpoint = endpoint
|
self.endpoint = endpoint
|
||||||
|
|
@ -65,7 +65,7 @@ class LiteLLMEmbeddingEngine(EmbeddingEngine):
|
||||||
self.provider = provider
|
self.provider = provider
|
||||||
self.model = model
|
self.model = model
|
||||||
self.dimensions = dimensions
|
self.dimensions = dimensions
|
||||||
self.max_tokens = max_tokens
|
self.max_completion_tokens = max_completion_tokens
|
||||||
self.tokenizer = self.get_tokenizer()
|
self.tokenizer = self.get_tokenizer()
|
||||||
self.retry_count = 0
|
self.retry_count = 0
|
||||||
|
|
||||||
|
|
@ -179,20 +179,29 @@ class LiteLLMEmbeddingEngine(EmbeddingEngine):
|
||||||
model = self.model.split("/")[-1]
|
model = self.model.split("/")[-1]
|
||||||
|
|
||||||
if "openai" in self.provider.lower():
|
if "openai" in self.provider.lower():
|
||||||
tokenizer = TikTokenTokenizer(model=model, max_tokens=self.max_tokens)
|
tokenizer = TikTokenTokenizer(
|
||||||
|
model=model, max_completion_tokens=self.max_completion_tokens
|
||||||
|
)
|
||||||
elif "gemini" in self.provider.lower():
|
elif "gemini" in self.provider.lower():
|
||||||
tokenizer = GeminiTokenizer(model=model, max_tokens=self.max_tokens)
|
tokenizer = GeminiTokenizer(
|
||||||
|
model=model, max_completion_tokens=self.max_completion_tokens
|
||||||
|
)
|
||||||
elif "mistral" in self.provider.lower():
|
elif "mistral" in self.provider.lower():
|
||||||
tokenizer = MistralTokenizer(model=model, max_tokens=self.max_tokens)
|
tokenizer = MistralTokenizer(
|
||||||
|
model=model, max_completion_tokens=self.max_completion_tokens
|
||||||
|
)
|
||||||
else:
|
else:
|
||||||
try:
|
try:
|
||||||
tokenizer = HuggingFaceTokenizer(
|
tokenizer = HuggingFaceTokenizer(
|
||||||
model=self.model.replace("hosted_vllm/", ""), max_tokens=self.max_tokens
|
model=self.model.replace("hosted_vllm/", ""),
|
||||||
|
max_completion_tokens=self.max_completion_tokens,
|
||||||
)
|
)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.warning(f"Could not get tokenizer from HuggingFace due to: {e}")
|
logger.warning(f"Could not get tokenizer from HuggingFace due to: {e}")
|
||||||
logger.info("Switching to TikToken default tokenizer.")
|
logger.info("Switching to TikToken default tokenizer.")
|
||||||
tokenizer = TikTokenTokenizer(model=None, max_tokens=self.max_tokens)
|
tokenizer = TikTokenTokenizer(
|
||||||
|
model=None, max_completion_tokens=self.max_completion_tokens
|
||||||
|
)
|
||||||
|
|
||||||
logger.debug(f"Tokenizer loaded for model: {self.model}")
|
logger.debug(f"Tokenizer loaded for model: {self.model}")
|
||||||
return tokenizer
|
return tokenizer
|
||||||
|
|
|
||||||
|
|
@ -30,7 +30,7 @@ class OllamaEmbeddingEngine(EmbeddingEngine):
|
||||||
Instance variables:
|
Instance variables:
|
||||||
- model
|
- model
|
||||||
- dimensions
|
- dimensions
|
||||||
- max_tokens
|
- max_completion_tokens
|
||||||
- endpoint
|
- endpoint
|
||||||
- mock
|
- mock
|
||||||
- huggingface_tokenizer_name
|
- huggingface_tokenizer_name
|
||||||
|
|
@ -39,7 +39,7 @@ class OllamaEmbeddingEngine(EmbeddingEngine):
|
||||||
|
|
||||||
model: str
|
model: str
|
||||||
dimensions: int
|
dimensions: int
|
||||||
max_tokens: int
|
max_completion_tokens: int
|
||||||
endpoint: str
|
endpoint: str
|
||||||
mock: bool
|
mock: bool
|
||||||
huggingface_tokenizer_name: str
|
huggingface_tokenizer_name: str
|
||||||
|
|
@ -50,13 +50,13 @@ class OllamaEmbeddingEngine(EmbeddingEngine):
|
||||||
self,
|
self,
|
||||||
model: Optional[str] = "avr/sfr-embedding-mistral:latest",
|
model: Optional[str] = "avr/sfr-embedding-mistral:latest",
|
||||||
dimensions: Optional[int] = 1024,
|
dimensions: Optional[int] = 1024,
|
||||||
max_tokens: int = 512,
|
max_completion_tokens: int = 512,
|
||||||
endpoint: Optional[str] = "http://localhost:11434/api/embeddings",
|
endpoint: Optional[str] = "http://localhost:11434/api/embeddings",
|
||||||
huggingface_tokenizer: str = "Salesforce/SFR-Embedding-Mistral",
|
huggingface_tokenizer: str = "Salesforce/SFR-Embedding-Mistral",
|
||||||
):
|
):
|
||||||
self.model = model
|
self.model = model
|
||||||
self.dimensions = dimensions
|
self.dimensions = dimensions
|
||||||
self.max_tokens = max_tokens
|
self.max_completion_tokens = max_completion_tokens
|
||||||
self.endpoint = endpoint
|
self.endpoint = endpoint
|
||||||
self.huggingface_tokenizer_name = huggingface_tokenizer
|
self.huggingface_tokenizer_name = huggingface_tokenizer
|
||||||
self.tokenizer = self.get_tokenizer()
|
self.tokenizer = self.get_tokenizer()
|
||||||
|
|
@ -132,7 +132,7 @@ class OllamaEmbeddingEngine(EmbeddingEngine):
|
||||||
"""
|
"""
|
||||||
logger.debug("Loading HuggingfaceTokenizer for OllamaEmbeddingEngine...")
|
logger.debug("Loading HuggingfaceTokenizer for OllamaEmbeddingEngine...")
|
||||||
tokenizer = HuggingFaceTokenizer(
|
tokenizer = HuggingFaceTokenizer(
|
||||||
model=self.huggingface_tokenizer_name, max_tokens=self.max_tokens
|
model=self.huggingface_tokenizer_name, max_completion_tokens=self.max_completion_tokens
|
||||||
)
|
)
|
||||||
logger.debug("Tokenizer loaded for OllamaEmbeddingEngine")
|
logger.debug("Tokenizer loaded for OllamaEmbeddingEngine")
|
||||||
return tokenizer
|
return tokenizer
|
||||||
|
|
|
||||||
|
|
@ -18,7 +18,7 @@ class EmbeddingConfig(BaseSettings):
|
||||||
embedding_endpoint: Optional[str] = None
|
embedding_endpoint: Optional[str] = None
|
||||||
embedding_api_key: Optional[str] = None
|
embedding_api_key: Optional[str] = None
|
||||||
embedding_api_version: Optional[str] = None
|
embedding_api_version: Optional[str] = None
|
||||||
embedding_max_tokens: Optional[int] = 8191
|
embedding_max_completion_tokens: Optional[int] = 8191
|
||||||
huggingface_tokenizer: Optional[str] = None
|
huggingface_tokenizer: Optional[str] = None
|
||||||
model_config = SettingsConfigDict(env_file=".env", extra="allow")
|
model_config = SettingsConfigDict(env_file=".env", extra="allow")
|
||||||
|
|
||||||
|
|
@ -38,7 +38,7 @@ class EmbeddingConfig(BaseSettings):
|
||||||
"embedding_endpoint": self.embedding_endpoint,
|
"embedding_endpoint": self.embedding_endpoint,
|
||||||
"embedding_api_key": self.embedding_api_key,
|
"embedding_api_key": self.embedding_api_key,
|
||||||
"embedding_api_version": self.embedding_api_version,
|
"embedding_api_version": self.embedding_api_version,
|
||||||
"embedding_max_tokens": self.embedding_max_tokens,
|
"embedding_max_completion_tokens": self.embedding_max_completion_tokens,
|
||||||
"huggingface_tokenizer": self.huggingface_tokenizer,
|
"huggingface_tokenizer": self.huggingface_tokenizer,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -27,7 +27,7 @@ def get_embedding_engine() -> EmbeddingEngine:
|
||||||
config.embedding_provider,
|
config.embedding_provider,
|
||||||
config.embedding_model,
|
config.embedding_model,
|
||||||
config.embedding_dimensions,
|
config.embedding_dimensions,
|
||||||
config.embedding_max_tokens,
|
config.embedding_max_completion_tokens,
|
||||||
config.embedding_endpoint,
|
config.embedding_endpoint,
|
||||||
config.embedding_api_key,
|
config.embedding_api_key,
|
||||||
config.embedding_api_version,
|
config.embedding_api_version,
|
||||||
|
|
@ -41,7 +41,7 @@ def create_embedding_engine(
|
||||||
embedding_provider,
|
embedding_provider,
|
||||||
embedding_model,
|
embedding_model,
|
||||||
embedding_dimensions,
|
embedding_dimensions,
|
||||||
embedding_max_tokens,
|
embedding_max_completion_tokens,
|
||||||
embedding_endpoint,
|
embedding_endpoint,
|
||||||
embedding_api_key,
|
embedding_api_key,
|
||||||
embedding_api_version,
|
embedding_api_version,
|
||||||
|
|
@ -58,7 +58,7 @@ def create_embedding_engine(
|
||||||
'ollama', or another supported provider.
|
'ollama', or another supported provider.
|
||||||
- embedding_model: The model to be used for the embedding engine.
|
- embedding_model: The model to be used for the embedding engine.
|
||||||
- embedding_dimensions: The number of dimensions for the embeddings.
|
- embedding_dimensions: The number of dimensions for the embeddings.
|
||||||
- embedding_max_tokens: The maximum number of tokens for the embeddings.
|
- embedding_max_completion_tokens: The maximum number of tokens for the embeddings.
|
||||||
- embedding_endpoint: The endpoint for the embedding service, relevant for certain
|
- embedding_endpoint: The endpoint for the embedding service, relevant for certain
|
||||||
providers.
|
providers.
|
||||||
- embedding_api_key: API key to authenticate with the embedding service, if
|
- embedding_api_key: API key to authenticate with the embedding service, if
|
||||||
|
|
@ -81,7 +81,7 @@ def create_embedding_engine(
|
||||||
return FastembedEmbeddingEngine(
|
return FastembedEmbeddingEngine(
|
||||||
model=embedding_model,
|
model=embedding_model,
|
||||||
dimensions=embedding_dimensions,
|
dimensions=embedding_dimensions,
|
||||||
max_tokens=embedding_max_tokens,
|
max_completion_tokens=embedding_max_completion_tokens,
|
||||||
)
|
)
|
||||||
|
|
||||||
if embedding_provider == "ollama":
|
if embedding_provider == "ollama":
|
||||||
|
|
@ -90,7 +90,7 @@ def create_embedding_engine(
|
||||||
return OllamaEmbeddingEngine(
|
return OllamaEmbeddingEngine(
|
||||||
model=embedding_model,
|
model=embedding_model,
|
||||||
dimensions=embedding_dimensions,
|
dimensions=embedding_dimensions,
|
||||||
max_tokens=embedding_max_tokens,
|
max_completion_tokens=embedding_max_completion_tokens,
|
||||||
endpoint=embedding_endpoint,
|
endpoint=embedding_endpoint,
|
||||||
huggingface_tokenizer=huggingface_tokenizer,
|
huggingface_tokenizer=huggingface_tokenizer,
|
||||||
)
|
)
|
||||||
|
|
@ -104,5 +104,5 @@ def create_embedding_engine(
|
||||||
api_version=embedding_api_version,
|
api_version=embedding_api_version,
|
||||||
model=embedding_model,
|
model=embedding_model,
|
||||||
dimensions=embedding_dimensions,
|
dimensions=embedding_dimensions,
|
||||||
max_tokens=embedding_max_tokens,
|
max_completion_tokens=embedding_max_completion_tokens,
|
||||||
)
|
)
|
||||||
|
|
|
||||||
|
|
@ -18,7 +18,7 @@ class LLMConfig(BaseSettings):
|
||||||
- llm_api_version
|
- llm_api_version
|
||||||
- llm_temperature
|
- llm_temperature
|
||||||
- llm_streaming
|
- llm_streaming
|
||||||
- llm_max_tokens
|
- llm_max_completion_tokens
|
||||||
- transcription_model
|
- transcription_model
|
||||||
- graph_prompt_path
|
- graph_prompt_path
|
||||||
- llm_rate_limit_enabled
|
- llm_rate_limit_enabled
|
||||||
|
|
@ -35,16 +35,16 @@ class LLMConfig(BaseSettings):
|
||||||
|
|
||||||
structured_output_framework: str = "instructor"
|
structured_output_framework: str = "instructor"
|
||||||
llm_provider: str = "openai"
|
llm_provider: str = "openai"
|
||||||
llm_model: str = "gpt-4o-mini"
|
llm_model: str = "gpt-5-mini"
|
||||||
llm_endpoint: str = ""
|
llm_endpoint: str = ""
|
||||||
llm_api_key: Optional[str] = None
|
llm_api_key: Optional[str] = None
|
||||||
llm_api_version: Optional[str] = None
|
llm_api_version: Optional[str] = None
|
||||||
llm_temperature: float = 0.0
|
llm_temperature: float = 0.0
|
||||||
llm_streaming: bool = False
|
llm_streaming: bool = False
|
||||||
llm_max_tokens: int = 16384
|
llm_max_completion_tokens: int = 16384
|
||||||
|
|
||||||
baml_llm_provider: str = "openai"
|
baml_llm_provider: str = "openai"
|
||||||
baml_llm_model: str = "gpt-4o-mini"
|
baml_llm_model: str = "gpt-5-mini"
|
||||||
baml_llm_endpoint: str = ""
|
baml_llm_endpoint: str = ""
|
||||||
baml_llm_api_key: Optional[str] = None
|
baml_llm_api_key: Optional[str] = None
|
||||||
baml_llm_temperature: float = 0.0
|
baml_llm_temperature: float = 0.0
|
||||||
|
|
@ -171,7 +171,7 @@ class LLMConfig(BaseSettings):
|
||||||
"api_version": self.llm_api_version,
|
"api_version": self.llm_api_version,
|
||||||
"temperature": self.llm_temperature,
|
"temperature": self.llm_temperature,
|
||||||
"streaming": self.llm_streaming,
|
"streaming": self.llm_streaming,
|
||||||
"max_tokens": self.llm_max_tokens,
|
"max_completion_tokens": self.llm_max_completion_tokens,
|
||||||
"transcription_model": self.transcription_model,
|
"transcription_model": self.transcription_model,
|
||||||
"graph_prompt_path": self.graph_prompt_path,
|
"graph_prompt_path": self.graph_prompt_path,
|
||||||
"rate_limit_enabled": self.llm_rate_limit_enabled,
|
"rate_limit_enabled": self.llm_rate_limit_enabled,
|
||||||
|
|
|
||||||
|
|
@ -23,7 +23,7 @@ class AnthropicAdapter(LLMInterface):
|
||||||
name = "Anthropic"
|
name = "Anthropic"
|
||||||
model: str
|
model: str
|
||||||
|
|
||||||
def __init__(self, max_tokens: int, model: str = None):
|
def __init__(self, max_completion_tokens: int, model: str = None):
|
||||||
import anthropic
|
import anthropic
|
||||||
|
|
||||||
self.aclient = instructor.patch(
|
self.aclient = instructor.patch(
|
||||||
|
|
@ -31,7 +31,7 @@ class AnthropicAdapter(LLMInterface):
|
||||||
)
|
)
|
||||||
|
|
||||||
self.model = model
|
self.model = model
|
||||||
self.max_tokens = max_tokens
|
self.max_completion_tokens = max_completion_tokens
|
||||||
|
|
||||||
@sleep_and_retry_async()
|
@sleep_and_retry_async()
|
||||||
@rate_limit_async
|
@rate_limit_async
|
||||||
|
|
@ -57,7 +57,7 @@ class AnthropicAdapter(LLMInterface):
|
||||||
|
|
||||||
return await self.aclient(
|
return await self.aclient(
|
||||||
model=self.model,
|
model=self.model,
|
||||||
max_tokens=4096,
|
max_completion_tokens=4096,
|
||||||
max_retries=5,
|
max_retries=5,
|
||||||
messages=[
|
messages=[
|
||||||
{
|
{
|
||||||
|
|
|
||||||
|
|
@ -34,7 +34,7 @@ class GeminiAdapter(LLMInterface):
|
||||||
self,
|
self,
|
||||||
api_key: str,
|
api_key: str,
|
||||||
model: str,
|
model: str,
|
||||||
max_tokens: int,
|
max_completion_tokens: int,
|
||||||
endpoint: Optional[str] = None,
|
endpoint: Optional[str] = None,
|
||||||
api_version: Optional[str] = None,
|
api_version: Optional[str] = None,
|
||||||
streaming: bool = False,
|
streaming: bool = False,
|
||||||
|
|
@ -44,7 +44,7 @@ class GeminiAdapter(LLMInterface):
|
||||||
self.endpoint = endpoint
|
self.endpoint = endpoint
|
||||||
self.api_version = api_version
|
self.api_version = api_version
|
||||||
self.streaming = streaming
|
self.streaming = streaming
|
||||||
self.max_tokens = max_tokens
|
self.max_completion_tokens = max_completion_tokens
|
||||||
|
|
||||||
@observe(as_type="generation")
|
@observe(as_type="generation")
|
||||||
@sleep_and_retry_async()
|
@sleep_and_retry_async()
|
||||||
|
|
@ -90,7 +90,7 @@ class GeminiAdapter(LLMInterface):
|
||||||
model=f"{self.model}",
|
model=f"{self.model}",
|
||||||
messages=messages,
|
messages=messages,
|
||||||
api_key=self.api_key,
|
api_key=self.api_key,
|
||||||
max_tokens=self.max_tokens,
|
max_completion_tokens=self.max_completion_tokens,
|
||||||
temperature=0.1,
|
temperature=0.1,
|
||||||
response_format=response_schema,
|
response_format=response_schema,
|
||||||
timeout=100,
|
timeout=100,
|
||||||
|
|
|
||||||
|
|
@ -41,7 +41,7 @@ class GenericAPIAdapter(LLMInterface):
|
||||||
api_key: str,
|
api_key: str,
|
||||||
model: str,
|
model: str,
|
||||||
name: str,
|
name: str,
|
||||||
max_tokens: int,
|
max_completion_tokens: int,
|
||||||
fallback_model: str = None,
|
fallback_model: str = None,
|
||||||
fallback_api_key: str = None,
|
fallback_api_key: str = None,
|
||||||
fallback_endpoint: str = None,
|
fallback_endpoint: str = None,
|
||||||
|
|
@ -50,7 +50,7 @@ class GenericAPIAdapter(LLMInterface):
|
||||||
self.model = model
|
self.model = model
|
||||||
self.api_key = api_key
|
self.api_key = api_key
|
||||||
self.endpoint = endpoint
|
self.endpoint = endpoint
|
||||||
self.max_tokens = max_tokens
|
self.max_completion_tokens = max_completion_tokens
|
||||||
|
|
||||||
self.fallback_model = fallback_model
|
self.fallback_model = fallback_model
|
||||||
self.fallback_api_key = fallback_api_key
|
self.fallback_api_key = fallback_api_key
|
||||||
|
|
|
||||||
|
|
@ -54,11 +54,15 @@ def get_llm_client():
|
||||||
# Check if max_token value is defined in liteLLM for given model
|
# Check if max_token value is defined in liteLLM for given model
|
||||||
# if not use value from cognee configuration
|
# if not use value from cognee configuration
|
||||||
from cognee.infrastructure.llm.utils import (
|
from cognee.infrastructure.llm.utils import (
|
||||||
get_model_max_tokens,
|
get_model_max_completion_tokens,
|
||||||
) # imported here to avoid circular imports
|
) # imported here to avoid circular imports
|
||||||
|
|
||||||
model_max_tokens = get_model_max_tokens(llm_config.llm_model)
|
model_max_completion_tokens = get_model_max_completion_tokens(llm_config.llm_model)
|
||||||
max_tokens = model_max_tokens if model_max_tokens else llm_config.llm_max_tokens
|
max_completion_tokens = (
|
||||||
|
model_max_completion_tokens
|
||||||
|
if model_max_completion_tokens
|
||||||
|
else llm_config.llm_max_completion_tokens
|
||||||
|
)
|
||||||
|
|
||||||
if provider == LLMProvider.OPENAI:
|
if provider == LLMProvider.OPENAI:
|
||||||
if llm_config.llm_api_key is None:
|
if llm_config.llm_api_key is None:
|
||||||
|
|
@ -74,7 +78,7 @@ def get_llm_client():
|
||||||
api_version=llm_config.llm_api_version,
|
api_version=llm_config.llm_api_version,
|
||||||
model=llm_config.llm_model,
|
model=llm_config.llm_model,
|
||||||
transcription_model=llm_config.transcription_model,
|
transcription_model=llm_config.transcription_model,
|
||||||
max_tokens=max_tokens,
|
max_completion_tokens=max_completion_tokens,
|
||||||
streaming=llm_config.llm_streaming,
|
streaming=llm_config.llm_streaming,
|
||||||
fallback_api_key=llm_config.fallback_api_key,
|
fallback_api_key=llm_config.fallback_api_key,
|
||||||
fallback_endpoint=llm_config.fallback_endpoint,
|
fallback_endpoint=llm_config.fallback_endpoint,
|
||||||
|
|
@ -94,7 +98,7 @@ def get_llm_client():
|
||||||
llm_config.llm_api_key,
|
llm_config.llm_api_key,
|
||||||
llm_config.llm_model,
|
llm_config.llm_model,
|
||||||
"Ollama",
|
"Ollama",
|
||||||
max_tokens=max_tokens,
|
max_completion_tokens=max_completion_tokens,
|
||||||
)
|
)
|
||||||
|
|
||||||
elif provider == LLMProvider.ANTHROPIC:
|
elif provider == LLMProvider.ANTHROPIC:
|
||||||
|
|
@ -102,7 +106,9 @@ def get_llm_client():
|
||||||
AnthropicAdapter,
|
AnthropicAdapter,
|
||||||
)
|
)
|
||||||
|
|
||||||
return AnthropicAdapter(max_tokens=max_tokens, model=llm_config.llm_model)
|
return AnthropicAdapter(
|
||||||
|
max_completion_tokens=max_completion_tokens, model=llm_config.llm_model
|
||||||
|
)
|
||||||
|
|
||||||
elif provider == LLMProvider.CUSTOM:
|
elif provider == LLMProvider.CUSTOM:
|
||||||
if llm_config.llm_api_key is None:
|
if llm_config.llm_api_key is None:
|
||||||
|
|
@ -117,7 +123,7 @@ def get_llm_client():
|
||||||
llm_config.llm_api_key,
|
llm_config.llm_api_key,
|
||||||
llm_config.llm_model,
|
llm_config.llm_model,
|
||||||
"Custom",
|
"Custom",
|
||||||
max_tokens=max_tokens,
|
max_completion_tokens=max_completion_tokens,
|
||||||
fallback_api_key=llm_config.fallback_api_key,
|
fallback_api_key=llm_config.fallback_api_key,
|
||||||
fallback_endpoint=llm_config.fallback_endpoint,
|
fallback_endpoint=llm_config.fallback_endpoint,
|
||||||
fallback_model=llm_config.fallback_model,
|
fallback_model=llm_config.fallback_model,
|
||||||
|
|
@ -134,7 +140,7 @@ def get_llm_client():
|
||||||
return GeminiAdapter(
|
return GeminiAdapter(
|
||||||
api_key=llm_config.llm_api_key,
|
api_key=llm_config.llm_api_key,
|
||||||
model=llm_config.llm_model,
|
model=llm_config.llm_model,
|
||||||
max_tokens=max_tokens,
|
max_completion_tokens=max_completion_tokens,
|
||||||
endpoint=llm_config.llm_endpoint,
|
endpoint=llm_config.llm_endpoint,
|
||||||
api_version=llm_config.llm_api_version,
|
api_version=llm_config.llm_api_version,
|
||||||
streaming=llm_config.llm_streaming,
|
streaming=llm_config.llm_streaming,
|
||||||
|
|
|
||||||
|
|
@ -30,16 +30,18 @@ class OllamaAPIAdapter(LLMInterface):
|
||||||
- model
|
- model
|
||||||
- api_key
|
- api_key
|
||||||
- endpoint
|
- endpoint
|
||||||
- max_tokens
|
- max_completion_tokens
|
||||||
- aclient
|
- aclient
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(self, endpoint: str, api_key: str, model: str, name: str, max_tokens: int):
|
def __init__(
|
||||||
|
self, endpoint: str, api_key: str, model: str, name: str, max_completion_tokens: int
|
||||||
|
):
|
||||||
self.name = name
|
self.name = name
|
||||||
self.model = model
|
self.model = model
|
||||||
self.api_key = api_key
|
self.api_key = api_key
|
||||||
self.endpoint = endpoint
|
self.endpoint = endpoint
|
||||||
self.max_tokens = max_tokens
|
self.max_completion_tokens = max_completion_tokens
|
||||||
|
|
||||||
self.aclient = instructor.from_openai(
|
self.aclient = instructor.from_openai(
|
||||||
OpenAI(base_url=self.endpoint, api_key=self.api_key), mode=instructor.Mode.JSON
|
OpenAI(base_url=self.endpoint, api_key=self.api_key), mode=instructor.Mode.JSON
|
||||||
|
|
@ -159,7 +161,7 @@ class OllamaAPIAdapter(LLMInterface):
|
||||||
],
|
],
|
||||||
}
|
}
|
||||||
],
|
],
|
||||||
max_tokens=300,
|
max_completion_tokens=300,
|
||||||
)
|
)
|
||||||
|
|
||||||
# Ensure response is valid before accessing .choices[0].message.content
|
# Ensure response is valid before accessing .choices[0].message.content
|
||||||
|
|
|
||||||
|
|
@ -64,7 +64,7 @@ class OpenAIAdapter(LLMInterface):
|
||||||
api_version: str,
|
api_version: str,
|
||||||
model: str,
|
model: str,
|
||||||
transcription_model: str,
|
transcription_model: str,
|
||||||
max_tokens: int,
|
max_completion_tokens: int,
|
||||||
streaming: bool = False,
|
streaming: bool = False,
|
||||||
fallback_model: str = None,
|
fallback_model: str = None,
|
||||||
fallback_api_key: str = None,
|
fallback_api_key: str = None,
|
||||||
|
|
@ -77,7 +77,7 @@ class OpenAIAdapter(LLMInterface):
|
||||||
self.api_key = api_key
|
self.api_key = api_key
|
||||||
self.endpoint = endpoint
|
self.endpoint = endpoint
|
||||||
self.api_version = api_version
|
self.api_version = api_version
|
||||||
self.max_tokens = max_tokens
|
self.max_completion_tokens = max_completion_tokens
|
||||||
self.streaming = streaming
|
self.streaming = streaming
|
||||||
|
|
||||||
self.fallback_model = fallback_model
|
self.fallback_model = fallback_model
|
||||||
|
|
@ -301,7 +301,7 @@ class OpenAIAdapter(LLMInterface):
|
||||||
api_key=self.api_key,
|
api_key=self.api_key,
|
||||||
api_base=self.endpoint,
|
api_base=self.endpoint,
|
||||||
api_version=self.api_version,
|
api_version=self.api_version,
|
||||||
max_tokens=300,
|
max_completion_tokens=300,
|
||||||
max_retries=self.MAX_RETRIES,
|
max_retries=self.MAX_RETRIES,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -17,10 +17,10 @@ class GeminiTokenizer(TokenizerInterface):
|
||||||
def __init__(
|
def __init__(
|
||||||
self,
|
self,
|
||||||
model: str,
|
model: str,
|
||||||
max_tokens: int = 3072,
|
max_completion_tokens: int = 3072,
|
||||||
):
|
):
|
||||||
self.model = model
|
self.model = model
|
||||||
self.max_tokens = max_tokens
|
self.max_completion_tokens = max_completion_tokens
|
||||||
|
|
||||||
# Get LLM API key from config
|
# Get LLM API key from config
|
||||||
from cognee.infrastructure.databases.vector.embeddings.config import get_embedding_config
|
from cognee.infrastructure.databases.vector.embeddings.config import get_embedding_config
|
||||||
|
|
|
||||||
|
|
@ -14,17 +14,17 @@ class HuggingFaceTokenizer(TokenizerInterface):
|
||||||
|
|
||||||
Instance variables include:
|
Instance variables include:
|
||||||
- model: str
|
- model: str
|
||||||
- max_tokens: int
|
- max_completion_tokens: int
|
||||||
- tokenizer: AutoTokenizer
|
- tokenizer: AutoTokenizer
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(
|
def __init__(
|
||||||
self,
|
self,
|
||||||
model: str,
|
model: str,
|
||||||
max_tokens: int = 512,
|
max_completion_tokens: int = 512,
|
||||||
):
|
):
|
||||||
self.model = model
|
self.model = model
|
||||||
self.max_tokens = max_tokens
|
self.max_completion_tokens = max_completion_tokens
|
||||||
|
|
||||||
# Import here to make it an optional dependency
|
# Import here to make it an optional dependency
|
||||||
from transformers import AutoTokenizer
|
from transformers import AutoTokenizer
|
||||||
|
|
|
||||||
|
|
@ -16,17 +16,17 @@ class MistralTokenizer(TokenizerInterface):
|
||||||
|
|
||||||
Instance variables include:
|
Instance variables include:
|
||||||
- model: str
|
- model: str
|
||||||
- max_tokens: int
|
- max_completion_tokens: int
|
||||||
- tokenizer: MistralTokenizer
|
- tokenizer: MistralTokenizer
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(
|
def __init__(
|
||||||
self,
|
self,
|
||||||
model: str,
|
model: str,
|
||||||
max_tokens: int = 3072,
|
max_completion_tokens: int = 3072,
|
||||||
):
|
):
|
||||||
self.model = model
|
self.model = model
|
||||||
self.max_tokens = max_tokens
|
self.max_completion_tokens = max_completion_tokens
|
||||||
|
|
||||||
# Import here to make it an optional dependency
|
# Import here to make it an optional dependency
|
||||||
from mistral_common.tokens.tokenizers.mistral import MistralTokenizer
|
from mistral_common.tokens.tokenizers.mistral import MistralTokenizer
|
||||||
|
|
|
||||||
|
|
@ -13,10 +13,10 @@ class TikTokenTokenizer(TokenizerInterface):
|
||||||
def __init__(
|
def __init__(
|
||||||
self,
|
self,
|
||||||
model: Optional[str] = None,
|
model: Optional[str] = None,
|
||||||
max_tokens: int = 8191,
|
max_completion_tokens: int = 8191,
|
||||||
):
|
):
|
||||||
self.model = model
|
self.model = model
|
||||||
self.max_tokens = max_tokens
|
self.max_completion_tokens = max_completion_tokens
|
||||||
# Initialize TikToken for GPT based on model
|
# Initialize TikToken for GPT based on model
|
||||||
if model:
|
if model:
|
||||||
self.tokenizer = tiktoken.encoding_for_model(self.model)
|
self.tokenizer = tiktoken.encoding_for_model(self.model)
|
||||||
|
|
@ -93,9 +93,9 @@ class TikTokenTokenizer(TokenizerInterface):
|
||||||
num_tokens = len(self.tokenizer.encode(text))
|
num_tokens = len(self.tokenizer.encode(text))
|
||||||
return num_tokens
|
return num_tokens
|
||||||
|
|
||||||
def trim_text_to_max_tokens(self, text: str) -> str:
|
def trim_text_to_max_completion_tokens(self, text: str) -> str:
|
||||||
"""
|
"""
|
||||||
Trim the text so that the number of tokens does not exceed max_tokens.
|
Trim the text so that the number of tokens does not exceed max_completion_tokens.
|
||||||
|
|
||||||
Parameters:
|
Parameters:
|
||||||
-----------
|
-----------
|
||||||
|
|
@ -111,13 +111,13 @@ class TikTokenTokenizer(TokenizerInterface):
|
||||||
num_tokens = self.count_tokens(text)
|
num_tokens = self.count_tokens(text)
|
||||||
|
|
||||||
# If the number of tokens is within the limit, return the text as is
|
# If the number of tokens is within the limit, return the text as is
|
||||||
if num_tokens <= self.max_tokens:
|
if num_tokens <= self.max_completion_tokens:
|
||||||
return text
|
return text
|
||||||
|
|
||||||
# If the number exceeds the limit, trim the text
|
# If the number exceeds the limit, trim the text
|
||||||
# This is a simple trim, it may cut words in half; consider using word boundaries for a cleaner cut
|
# This is a simple trim, it may cut words in half; consider using word boundaries for a cleaner cut
|
||||||
encoded_text = self.tokenizer.encode(text)
|
encoded_text = self.tokenizer.encode(text)
|
||||||
trimmed_encoded_text = encoded_text[: self.max_tokens]
|
trimmed_encoded_text = encoded_text[: self.max_completion_tokens]
|
||||||
# Decoding the trimmed text
|
# Decoding the trimmed text
|
||||||
trimmed_text = self.tokenizer.decode(trimmed_encoded_text)
|
trimmed_text = self.tokenizer.decode(trimmed_encoded_text)
|
||||||
return trimmed_text
|
return trimmed_text
|
||||||
|
|
|
||||||
|
|
@ -32,13 +32,13 @@ def get_max_chunk_tokens():
|
||||||
|
|
||||||
# We need to make sure chunk size won't take more than half of LLM max context token size
|
# We need to make sure chunk size won't take more than half of LLM max context token size
|
||||||
# but it also can't be bigger than the embedding engine max token size
|
# but it also can't be bigger than the embedding engine max token size
|
||||||
llm_cutoff_point = llm_client.max_tokens // 2 # Round down the division
|
llm_cutoff_point = llm_client.max_completion_tokens // 2 # Round down the division
|
||||||
max_chunk_tokens = min(embedding_engine.max_tokens, llm_cutoff_point)
|
max_chunk_tokens = min(embedding_engine.max_completion_tokens, llm_cutoff_point)
|
||||||
|
|
||||||
return max_chunk_tokens
|
return max_chunk_tokens
|
||||||
|
|
||||||
|
|
||||||
def get_model_max_tokens(model_name: str):
|
def get_model_max_completion_tokens(model_name: str):
|
||||||
"""
|
"""
|
||||||
Retrieve the maximum token limit for a specified model name if it exists.
|
Retrieve the maximum token limit for a specified model name if it exists.
|
||||||
|
|
||||||
|
|
@ -56,15 +56,15 @@ def get_model_max_tokens(model_name: str):
|
||||||
|
|
||||||
Number of max tokens of model, or None if model is unknown
|
Number of max tokens of model, or None if model is unknown
|
||||||
"""
|
"""
|
||||||
max_tokens = None
|
max_completion_tokens = None
|
||||||
|
|
||||||
if model_name in litellm.model_cost:
|
if model_name in litellm.model_cost:
|
||||||
max_tokens = litellm.model_cost[model_name]["max_tokens"]
|
max_completion_tokens = litellm.model_cost[model_name]["max_tokens"]
|
||||||
logger.debug(f"Max input tokens for {model_name}: {max_tokens}")
|
logger.debug(f"Max input tokens for {model_name}: {max_completion_tokens}")
|
||||||
else:
|
else:
|
||||||
logger.info("Model not found in LiteLLM's model_cost.")
|
logger.info("Model not found in LiteLLM's model_cost.")
|
||||||
|
|
||||||
return max_tokens
|
return max_completion_tokens
|
||||||
|
|
||||||
|
|
||||||
async def test_llm_connection():
|
async def test_llm_connection():
|
||||||
|
|
|
||||||
|
|
@ -88,8 +88,8 @@ def get_settings() -> SettingsDict:
|
||||||
"models": {
|
"models": {
|
||||||
"openai": [
|
"openai": [
|
||||||
{
|
{
|
||||||
"value": "gpt-4o-mini",
|
"value": "gpt-5-mini",
|
||||||
"label": "gpt-4o-mini",
|
"label": "gpt-5-mini",
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"value": "gpt-4o",
|
"value": "gpt-4o",
|
||||||
|
|
|
||||||
|
|
@ -121,7 +121,7 @@ async def main():
|
||||||
evaluation_metrics=["correctness", "EM", "f1"],
|
evaluation_metrics=["correctness", "EM", "f1"],
|
||||||
calculate_metrics=True,
|
calculate_metrics=True,
|
||||||
dashboard=True,
|
dashboard=True,
|
||||||
deepeval_model="gpt-4o-mini",
|
deepeval_model="gpt-5-mini",
|
||||||
).to_dict()
|
).to_dict()
|
||||||
|
|
||||||
logger.info(f"Starting evaluation of {len(json_file_paths)} JSON files")
|
logger.info(f"Starting evaluation of {len(json_file_paths)} JSON files")
|
||||||
|
|
|
||||||
|
|
@ -35,7 +35,7 @@ def _get_sources_from_corpus_json(corpus_file, limit=None, batch_size=None):
|
||||||
def create_ontology(
|
def create_ontology(
|
||||||
corpus_file="hotpot_50_corpus.json",
|
corpus_file="hotpot_50_corpus.json",
|
||||||
output_file="hotpot_qa_ontology.json",
|
output_file="hotpot_qa_ontology.json",
|
||||||
model_name="gpt-4o-mini",
|
model_name="gpt-5-mini",
|
||||||
):
|
):
|
||||||
"""Creates ontology from corpus data."""
|
"""Creates ontology from corpus data."""
|
||||||
print(f"Loading corpus from {corpus_file}...")
|
print(f"Loading corpus from {corpus_file}...")
|
||||||
|
|
@ -67,7 +67,7 @@ def create_knowledge_graph(
|
||||||
ontology_file="hotpot_qa_ontology.json",
|
ontology_file="hotpot_qa_ontology.json",
|
||||||
corpus_file="hotpot_50_corpus.json",
|
corpus_file="hotpot_50_corpus.json",
|
||||||
kg_name="hotpot_qa_kg",
|
kg_name="hotpot_qa_kg",
|
||||||
model_name="gpt-4o-mini",
|
model_name="gpt-5-mini",
|
||||||
host="127.0.0.1",
|
host="127.0.0.1",
|
||||||
port=6379,
|
port=6379,
|
||||||
username=None,
|
username=None,
|
||||||
|
|
@ -208,11 +208,11 @@ if __name__ == "__main__":
|
||||||
# Ontology parameters
|
# Ontology parameters
|
||||||
corpus_file: str = "hotpot_50_corpus.json"
|
corpus_file: str = "hotpot_50_corpus.json"
|
||||||
ontology_file: str = "hotpot_qa_ontology.json"
|
ontology_file: str = "hotpot_qa_ontology.json"
|
||||||
ontology_model: str = "gpt-4o-mini"
|
ontology_model: str = "gpt-5-mini"
|
||||||
|
|
||||||
# Knowledge graph parameters
|
# Knowledge graph parameters
|
||||||
kg_name: str = "hotpot_qa"
|
kg_name: str = "hotpot_qa"
|
||||||
kg_model: str = "gpt-4o-mini"
|
kg_model: str = "gpt-5-mini"
|
||||||
kg_host: str = "127.0.0.1"
|
kg_host: str = "127.0.0.1"
|
||||||
kg_port: int = 6379
|
kg_port: int = 6379
|
||||||
kg_recreate: bool = False
|
kg_recreate: bool = False
|
||||||
|
|
|
||||||
|
|
@ -50,7 +50,7 @@ async def search_graphiti(query: str, graphiti_client: Graphiti, num_results: in
|
||||||
|
|
||||||
async def answer_questions(
|
async def answer_questions(
|
||||||
graphiti: Graphiti,
|
graphiti: Graphiti,
|
||||||
model_name: str = "gpt-4o-mini",
|
model_name: str = "gpt-5-mini",
|
||||||
qa_pairs_file: str = "hotpot_50_qa_pairs.json",
|
qa_pairs_file: str = "hotpot_50_qa_pairs.json",
|
||||||
print_results: bool = True,
|
print_results: bool = True,
|
||||||
output_file: str = None,
|
output_file: str = None,
|
||||||
|
|
@ -173,7 +173,7 @@ if __name__ == "__main__":
|
||||||
corpus_limit: int = None # Limit number of documents to process
|
corpus_limit: int = None # Limit number of documents to process
|
||||||
|
|
||||||
# Model parameters
|
# Model parameters
|
||||||
model_name: str = "gpt-4o-mini"
|
model_name: str = "gpt-5-mini"
|
||||||
|
|
||||||
# QA parameters
|
# QA parameters
|
||||||
qa_pairs_file: str = "hotpot_50_qa_pairs.json"
|
qa_pairs_file: str = "hotpot_50_qa_pairs.json"
|
||||||
|
|
|
||||||
|
|
@ -41,7 +41,7 @@ def load_corpus_to_memory(
|
||||||
def answer_questions(
|
def answer_questions(
|
||||||
memory: Memory,
|
memory: Memory,
|
||||||
openai_client: OpenAI,
|
openai_client: OpenAI,
|
||||||
model_name: str = "gpt-4o-mini",
|
model_name: str = "gpt-5-mini",
|
||||||
user_id: str = "hotpot_qa_user",
|
user_id: str = "hotpot_qa_user",
|
||||||
qa_pairs_file: str = "hotpot_50_qa_pairs.json",
|
qa_pairs_file: str = "hotpot_50_qa_pairs.json",
|
||||||
print_results: bool = True,
|
print_results: bool = True,
|
||||||
|
|
@ -150,7 +150,7 @@ if __name__ == "__main__":
|
||||||
user_id: str = "hotpot_qa_user"
|
user_id: str = "hotpot_qa_user"
|
||||||
|
|
||||||
# Model parameters
|
# Model parameters
|
||||||
model_name: str = "gpt-4o-mini"
|
model_name: str = "gpt-5-mini"
|
||||||
|
|
||||||
# QA parameters
|
# QA parameters
|
||||||
qa_pairs_file: str = "hotpot_50_qa_pairs.json"
|
qa_pairs_file: str = "hotpot_50_qa_pairs.json"
|
||||||
|
|
|
||||||
|
|
@ -121,7 +121,7 @@ async def main():
|
||||||
evaluation_metrics=["correctness", "EM", "f1"],
|
evaluation_metrics=["correctness", "EM", "f1"],
|
||||||
calculate_metrics=True,
|
calculate_metrics=True,
|
||||||
dashboard=True,
|
dashboard=True,
|
||||||
deepeval_model="gpt-4o-mini",
|
deepeval_model="gpt-5-mini",
|
||||||
).to_dict()
|
).to_dict()
|
||||||
|
|
||||||
logger.info(f"Starting evaluation of {len(json_file_paths)} JSON files")
|
logger.info(f"Starting evaluation of {len(json_file_paths)} JSON files")
|
||||||
|
|
|
||||||
|
|
@ -26,7 +26,7 @@ class GraphitiConfig(QABenchmarkConfig):
|
||||||
db_password: str = os.getenv("NEO4J_PASSWORD")
|
db_password: str = os.getenv("NEO4J_PASSWORD")
|
||||||
|
|
||||||
# Model parameters
|
# Model parameters
|
||||||
# model_name: str = "gpt-4o-mini"
|
# model_name: str = "gpt-5-mini"
|
||||||
model_name: str = "gpt-4o"
|
model_name: str = "gpt-4o"
|
||||||
|
|
||||||
# Default results file
|
# Default results file
|
||||||
|
|
@ -43,7 +43,7 @@ class QABenchmarkGraphiti(QABenchmarkRAG):
|
||||||
|
|
||||||
async def initialize_rag(self) -> Any:
|
async def initialize_rag(self) -> Any:
|
||||||
"""Initialize Graphiti and LLM."""
|
"""Initialize Graphiti and LLM."""
|
||||||
llm_config = LLMConfig(model=self.config.model_name, max_tokens=65536)
|
llm_config = LLMConfig(model=self.config.model_name, max_completion_tokens=65536)
|
||||||
llm_client = OpenAIClient(config=llm_config)
|
llm_client = OpenAIClient(config=llm_config)
|
||||||
graphiti = Graphiti(
|
graphiti = Graphiti(
|
||||||
self.config.db_url,
|
self.config.db_url,
|
||||||
|
|
|
||||||
|
|
@ -19,7 +19,7 @@ class Mem0Config(QABenchmarkConfig):
|
||||||
user_id: str = "hotpot_qa_user"
|
user_id: str = "hotpot_qa_user"
|
||||||
|
|
||||||
# Model parameters
|
# Model parameters
|
||||||
# model_name: str = "gpt-4o-mini"
|
# model_name: str = "gpt-5-mini"
|
||||||
model_name: str = "gpt-4o"
|
model_name: str = "gpt-4o"
|
||||||
|
|
||||||
# Default results file
|
# Default results file
|
||||||
|
|
|
||||||
|
|
@ -3,7 +3,7 @@ import cognee
|
||||||
|
|
||||||
import os
|
import os
|
||||||
|
|
||||||
# By default cognee uses OpenAI's gpt-4o-mini LLM model
|
# By default cognee uses OpenAI's gpt-5-mini LLM model
|
||||||
# Provide your OpenAI LLM API KEY
|
# Provide your OpenAI LLM API KEY
|
||||||
os.environ["LLM_API_KEY"] = ""
|
os.environ["LLM_API_KEY"] = ""
|
||||||
|
|
||||||
|
|
|
||||||
26
notebooks/cognee_simple_demo.ipynb
vendored
26
notebooks/cognee_simple_demo.ipynb
vendored
|
|
@ -29,7 +29,7 @@
|
||||||
"id": "f51e92e9fdcf77b7",
|
"id": "f51e92e9fdcf77b7",
|
||||||
"metadata": {},
|
"metadata": {},
|
||||||
"source": [
|
"source": [
|
||||||
" By default cognee uses OpenAI's gpt-4o-mini LLM model.\n",
|
" By default cognee uses OpenAI's gpt-5-mini LLM model.\n",
|
||||||
"\n",
|
"\n",
|
||||||
" Provide your OpenAI LLM API KEY in the step bellow. Here's a guide on how to get your [OpenAI API key](https://help.openai.com/en/articles/4936850-where-do-i-find-my-openai-api-key)."
|
" Provide your OpenAI LLM API KEY in the step bellow. Here's a guide on how to get your [OpenAI API key](https://help.openai.com/en/articles/4936850-where-do-i-find-my-openai-api-key)."
|
||||||
]
|
]
|
||||||
|
|
@ -208,11 +208,11 @@
|
||||||
"\u001B[1mselected model name for cost calculation: openai/text-embedding-3-large\u001B[0m\u001B[92m20:16:56 - LiteLLM:INFO\u001B[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001B[1mselected model name for cost calculation: openai/text-embedding-3-large\u001B[0m\u001B[92m20:16:56 - LiteLLM:INFO\u001B[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001B[1mselected model name for cost calculation: openai/text-embedding-3-large\u001B[0m\u001B[92m20:16:56 - LiteLLM:INFO\u001B[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001B[1mselected model name for cost calculation: openai/text-embedding-3-large\u001B[0m\u001B[92m20:16:56 - LiteLLM:INFO\u001B[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001B[1mselected model name for cost calculation: openai/text-embedding-3-large\u001B[0m\u001B[92m20:16:57 - LiteLLM:INFO\u001B[0m: utils.py:3101 - \n",
|
"\u001B[1mselected model name for cost calculation: openai/text-embedding-3-large\u001B[0m\u001B[92m20:16:57 - LiteLLM:INFO\u001B[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001B[1m\n",
|
"\u001B[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001B[0m\u001B[92m20:16:59 - LiteLLM:INFO\u001B[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001B[0m\u001B[92m20:16:59 - LiteLLM:INFO\u001B[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001B[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001B[0m\u001B[92m20:16:59 - LiteLLM:INFO\u001B[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001B[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001B[0m\u001B[92m20:16:59 - LiteLLM:INFO\u001B[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001B[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001B[0m"
|
"\u001B[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001B[0m"
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
|
|
@ -264,11 +264,11 @@
|
||||||
"\u001B[1mselected model name for cost calculation: openai/text-embedding-3-large\u001B[0m\u001B[92m20:17:03 - LiteLLM:INFO\u001B[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001B[1mselected model name for cost calculation: openai/text-embedding-3-large\u001B[0m\u001B[92m20:17:03 - LiteLLM:INFO\u001B[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001B[1mselected model name for cost calculation: openai/text-embedding-3-large\u001B[0m\u001B[92m20:17:03 - LiteLLM:INFO\u001B[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001B[1mselected model name for cost calculation: openai/text-embedding-3-large\u001B[0m\u001B[92m20:17:03 - LiteLLM:INFO\u001B[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001B[1mselected model name for cost calculation: openai/text-embedding-3-large\u001B[0m\u001B[92m20:17:03 - LiteLLM:INFO\u001B[0m: utils.py:3101 - \n",
|
"\u001B[1mselected model name for cost calculation: openai/text-embedding-3-large\u001B[0m\u001B[92m20:17:03 - LiteLLM:INFO\u001B[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001B[1m\n",
|
"\u001B[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001B[0m\u001B[92m20:17:05 - LiteLLM:INFO\u001B[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001B[0m\u001B[92m20:17:05 - LiteLLM:INFO\u001B[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001B[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001B[0m\u001B[92m20:17:05 - LiteLLM:INFO\u001B[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001B[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001B[0m\u001B[92m20:17:05 - LiteLLM:INFO\u001B[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001B[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001B[0m"
|
"\u001B[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001B[0m"
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
|
|
@ -320,11 +320,11 @@
|
||||||
"\u001B[1mselected model name for cost calculation: openai/text-embedding-3-large\u001B[0m\u001B[92m20:17:06 - LiteLLM:INFO\u001B[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001B[1mselected model name for cost calculation: openai/text-embedding-3-large\u001B[0m\u001B[92m20:17:06 - LiteLLM:INFO\u001B[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001B[1mselected model name for cost calculation: openai/text-embedding-3-large\u001B[0m\u001B[92m20:17:07 - LiteLLM:INFO\u001B[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001B[1mselected model name for cost calculation: openai/text-embedding-3-large\u001B[0m\u001B[92m20:17:07 - LiteLLM:INFO\u001B[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001B[1mselected model name for cost calculation: openai/text-embedding-3-large\u001B[0m\u001B[92m20:17:07 - LiteLLM:INFO\u001B[0m: utils.py:3101 - \n",
|
"\u001B[1mselected model name for cost calculation: openai/text-embedding-3-large\u001B[0m\u001B[92m20:17:07 - LiteLLM:INFO\u001B[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001B[1m\n",
|
"\u001B[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001B[0m\u001B[92m20:17:08 - LiteLLM:INFO\u001B[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001B[0m\u001B[92m20:17:08 - LiteLLM:INFO\u001B[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001B[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001B[0m\u001B[92m20:17:08 - LiteLLM:INFO\u001B[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001B[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001B[0m\u001B[92m20:17:08 - LiteLLM:INFO\u001B[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001B[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001B[0m"
|
"\u001B[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001B[0m"
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
|
|
|
||||||
266
notebooks/ontology_demo.ipynb
vendored
266
notebooks/ontology_demo.ipynb
vendored
|
|
@ -192,12 +192,12 @@
|
||||||
"text": [
|
"text": [
|
||||||
"\n",
|
"\n",
|
||||||
"\u001b[1mLangfuse client is disabled since no public_key was provided as a parameter or environment variable 'LANGFUSE_PUBLIC_KEY'. See our docs: https://langfuse.com/docs/sdk/python/low-level-sdk#initialize-client\u001b[0m\u001b[92m20:23:36 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"\u001b[1mLangfuse client is disabled since no public_key was provided as a parameter or environment variable 'LANGFUSE_PUBLIC_KEY'. See our docs: https://langfuse.com/docs/sdk/python/low-level-sdk#initialize-client\u001b[0m\u001b[92m20:23:36 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:23:37 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:23:37 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\n",
|
||||||
"\u001b[1mEmbeddingRateLimiter initialized: enabled=False, requests_limit=60, interval_seconds=60\u001b[0m\u001b[92m20:23:37 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mEmbeddingRateLimiter initialized: enabled=False, requests_limit=60, interval_seconds=60\u001b[0m\u001b[92m20:23:37 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:23:37 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:23:37 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\n",
|
||||||
"\u001b[2m2025-06-18T18:23:38.051934\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mPipeline run started: `2bec40b8-e3d1-54ab-bfc5-eb5d4695ce63`\u001b[0m [\u001b[0m\u001b[1m\u001b[34mrun_tasks(tasks: [Task], data)\u001b[0m]\u001b[0m\n",
|
"\u001b[2m2025-06-18T18:23:38.051934\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mPipeline run started: `2bec40b8-e3d1-54ab-bfc5-eb5d4695ce63`\u001b[0m [\u001b[0m\u001b[1m\u001b[34mrun_tasks(tasks: [Task], data)\u001b[0m]\u001b[0m\n",
|
||||||
"\u001b[2m2025-06-18T18:23:38.052396\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mCoroutine task started: `resolve_data_directories`\u001b[0m [\u001b[0m\u001b[1m\u001b[34mrun_tasks_base\u001b[0m]\u001b[0m\n",
|
"\u001b[2m2025-06-18T18:23:38.052396\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mCoroutine task started: `resolve_data_directories`\u001b[0m [\u001b[0m\u001b[1m\u001b[34mrun_tasks_base\u001b[0m]\u001b[0m\n",
|
||||||
|
|
@ -225,36 +225,36 @@
|
||||||
"\u001b[2m2025-06-18T18:23:38.513906\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mCoroutine task started: `check_permissions_on_dataset`\u001b[0m [\u001b[0m\u001b[1m\u001b[34mrun_tasks_base\u001b[0m]\u001b[0m\n",
|
"\u001b[2m2025-06-18T18:23:38.513906\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mCoroutine task started: `check_permissions_on_dataset`\u001b[0m [\u001b[0m\u001b[1m\u001b[34mrun_tasks_base\u001b[0m]\u001b[0m\n",
|
||||||
"\u001b[2m2025-06-18T18:23:38.517146\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mAsync Generator task started: `extract_chunks_from_documents`\u001b[0m [\u001b[0m\u001b[1m\u001b[34mrun_tasks_base\u001b[0m]\u001b[0m\n",
|
"\u001b[2m2025-06-18T18:23:38.517146\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mAsync Generator task started: `extract_chunks_from_documents`\u001b[0m [\u001b[0m\u001b[1m\u001b[34mrun_tasks_base\u001b[0m]\u001b[0m\n",
|
||||||
"\u001b[2m2025-06-18T18:23:38.799196\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mCoroutine task started: `extract_graph_from_data`\u001b[0m [\u001b[0m\u001b[1m\u001b[34mrun_tasks_base\u001b[0m]\u001b[0m\u001b[92m20:23:38 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"\u001b[2m2025-06-18T18:23:38.799196\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mCoroutine task started: `extract_graph_from_data`\u001b[0m [\u001b[0m\u001b[1m\u001b[34mrun_tasks_base\u001b[0m]\u001b[0m\u001b[92m20:23:38 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:23:38 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:23:38 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:23:38 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:23:38 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:23:38 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:23:38 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:23:38 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:23:38 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:23:38 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:23:38 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:23:49 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:23:49 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:23:49 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:23:49 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:23:53 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:23:53 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:23:53 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:23:53 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:23:53 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:23:53 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:23:53 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:23:53 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:23:56 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:23:56 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:23:56 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:23:56 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:23:57 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:23:57 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:23:57 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:23:57 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:23:59 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:23:59 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:23:59 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:23:59 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\n",
|
||||||
"\u001b[2m2025-06-18T18:23:59.600042\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mNo close match found for 'concept' in category 'classes'\u001b[0m [\u001b[0m\u001b[1m\u001b[34mOntologyAdapter\u001b[0m]\u001b[0m\n",
|
"\u001b[2m2025-06-18T18:23:59.600042\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mNo close match found for 'concept' in category 'classes'\u001b[0m [\u001b[0m\u001b[1m\u001b[34mOntologyAdapter\u001b[0m]\u001b[0m\n",
|
||||||
"\u001b[2m2025-06-18T18:23:59.600534\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mNo close match found for 'coffee consumption' in category 'individuals'\u001b[0m [\u001b[0m\u001b[1m\u001b[34mOntologyAdapter\u001b[0m]\u001b[0m\n",
|
"\u001b[2m2025-06-18T18:23:59.600534\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mNo close match found for 'coffee consumption' in category 'individuals'\u001b[0m [\u001b[0m\u001b[1m\u001b[34mOntologyAdapter\u001b[0m]\u001b[0m\n",
|
||||||
"\u001b[2m2025-06-18T18:23:59.600963\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mNo close match found for 'cardiovascular health' in category 'individuals'\u001b[0m [\u001b[0m\u001b[1m\u001b[34mOntologyAdapter\u001b[0m]\u001b[0m\n",
|
"\u001b[2m2025-06-18T18:23:59.600963\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mNo close match found for 'cardiovascular health' in category 'individuals'\u001b[0m [\u001b[0m\u001b[1m\u001b[34mOntologyAdapter\u001b[0m]\u001b[0m\n",
|
||||||
|
|
@ -336,36 +336,36 @@
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:04 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:04 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\n",
|
||||||
"\u001b[2m2025-06-18T18:24:04.634561\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mCoroutine task started: `summarize_text`\u001b[0m [\u001b[0m\u001b[1m\u001b[34mrun_tasks_base\u001b[0m]\u001b[0m\u001b[92m20:24:04 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"\u001b[2m2025-06-18T18:24:04.634561\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mCoroutine task started: `summarize_text`\u001b[0m [\u001b[0m\u001b[1m\u001b[34mrun_tasks_base\u001b[0m]\u001b[0m\u001b[92m20:24:04 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:24:04 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:24:04 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:24:04 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:24:04 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:24:04 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:24:04 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:24:04 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:24:04 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:24:04 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:24:04 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:24:06 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:24:06 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:24:06 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:24:06 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:24:08 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:24:08 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:24:08 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:24:08 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:24:08 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:24:08 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:24:08 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:24:08 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:24:08 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:24:08 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:24:08 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:24:08 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:24:10 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:24:10 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:24:10 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:24:10 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:24:10 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:24:10 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:24:10 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:24:10 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\n",
|
||||||
"\u001b[2m2025-06-18T18:24:10.329954\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mCoroutine task started: `add_data_points`\u001b[0m [\u001b[0m\u001b[1m\u001b[34mrun_tasks_base\u001b[0m]\u001b[0m\u001b[92m20:24:11 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[2m2025-06-18T18:24:10.329954\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mCoroutine task started: `add_data_points`\u001b[0m [\u001b[0m\u001b[1m\u001b[34mrun_tasks_base\u001b[0m]\u001b[0m\u001b[92m20:24:11 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:12 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:12 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:13 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:13 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
|
|
@ -403,11 +403,11 @@
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:16 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:16 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:17 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:17 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:17 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:17 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:24:19 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:24:19 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:24:19 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:24:19 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:24:19 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:24:19 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:20 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:20 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:20 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:20 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:20 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:20 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
|
|
@ -431,11 +431,11 @@
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:21 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:21 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:21 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:21 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:21 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:21 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:24:23 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:24:23 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:24:23 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:24:23 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:24:24 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:24:24 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:24 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:24 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:24 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:24 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:24 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:24 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
|
|
@ -459,11 +459,11 @@
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:25 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:25 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:26 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:26 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:26 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:26 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:24:29 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:24:29 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:24:29 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:24:29 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:24:30 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:24:30 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:30 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:30 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:30 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:30 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:30 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:30 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
|
|
@ -487,11 +487,11 @@
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:30 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:30 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:31 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:31 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:31 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:31 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:24:33 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:24:33 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:24:33 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:24:33 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m"
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m"
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
|
|
@ -580,36 +580,36 @@
|
||||||
"\u001b[2m2025-06-18T18:24:33.516136\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mCoroutine task started: `check_permissions_on_dataset`\u001b[0m [\u001b[0m\u001b[1m\u001b[34mrun_tasks_base\u001b[0m]\u001b[0m\n",
|
"\u001b[2m2025-06-18T18:24:33.516136\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mCoroutine task started: `check_permissions_on_dataset`\u001b[0m [\u001b[0m\u001b[1m\u001b[34mrun_tasks_base\u001b[0m]\u001b[0m\n",
|
||||||
"\u001b[2m2025-06-18T18:24:33.519493\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mAsync Generator task started: `extract_chunks_from_documents`\u001b[0m [\u001b[0m\u001b[1m\u001b[34mrun_tasks_base\u001b[0m]\u001b[0m\n",
|
"\u001b[2m2025-06-18T18:24:33.519493\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mAsync Generator task started: `extract_chunks_from_documents`\u001b[0m [\u001b[0m\u001b[1m\u001b[34mrun_tasks_base\u001b[0m]\u001b[0m\n",
|
||||||
"\u001b[2m2025-06-18T18:24:33.807889\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mCoroutine task started: `extract_graph_from_data`\u001b[0m [\u001b[0m\u001b[1m\u001b[34mrun_tasks_base\u001b[0m]\u001b[0m\u001b[92m20:24:33 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"\u001b[2m2025-06-18T18:24:33.807889\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mCoroutine task started: `extract_graph_from_data`\u001b[0m [\u001b[0m\u001b[1m\u001b[34mrun_tasks_base\u001b[0m]\u001b[0m\u001b[92m20:24:33 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:24:33 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:24:33 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:24:33 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:24:33 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:24:33 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:24:33 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:24:33 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:24:33 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:24:33 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:24:33 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:24:42 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:24:42 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:24:42 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:24:42 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:24:43 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:24:43 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:24:43 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:24:43 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:24:46 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:24:46 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:24:46 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:24:46 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:24:52 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:24:52 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:24:52 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:24:52 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:24:53 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:24:53 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:24:53 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:24:53 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:24:53 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:24:53 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:24:53 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:24:53 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\n",
|
||||||
"\u001b[2m2025-06-18T18:24:53.860301\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mNo close match found for 'person' in category 'classes'\u001b[0m [\u001b[0m\u001b[1m\u001b[34mOntologyAdapter\u001b[0m]\u001b[0m\n",
|
"\u001b[2m2025-06-18T18:24:53.860301\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mNo close match found for 'person' in category 'classes'\u001b[0m [\u001b[0m\u001b[1m\u001b[34mOntologyAdapter\u001b[0m]\u001b[0m\n",
|
||||||
"\u001b[2m2025-06-18T18:24:53.860877\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mNo close match found for 'laura torres-collado' in category 'individuals'\u001b[0m [\u001b[0m\u001b[1m\u001b[34mOntologyAdapter\u001b[0m]\u001b[0m\n",
|
"\u001b[2m2025-06-18T18:24:53.860877\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mNo close match found for 'laura torres-collado' in category 'individuals'\u001b[0m [\u001b[0m\u001b[1m\u001b[34mOntologyAdapter\u001b[0m]\u001b[0m\n",
|
||||||
"\u001b[2m2025-06-18T18:24:53.861358\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mNo close match found for 'laura maría compañ-gabucio' in category 'individuals'\u001b[0m [\u001b[0m\u001b[1m\u001b[34mOntologyAdapter\u001b[0m]\u001b[0m\n",
|
"\u001b[2m2025-06-18T18:24:53.861358\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mNo close match found for 'laura maría compañ-gabucio' in category 'individuals'\u001b[0m [\u001b[0m\u001b[1m\u001b[34mOntologyAdapter\u001b[0m]\u001b[0m\n",
|
||||||
|
|
@ -723,36 +723,36 @@
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:58 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:24:58 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\n",
|
||||||
"\u001b[2m2025-06-18T18:24:58.280117\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mCoroutine task started: `summarize_text`\u001b[0m [\u001b[0m\u001b[1m\u001b[34mrun_tasks_base\u001b[0m]\u001b[0m\u001b[92m20:24:58 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"\u001b[2m2025-06-18T18:24:58.280117\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mCoroutine task started: `summarize_text`\u001b[0m [\u001b[0m\u001b[1m\u001b[34mrun_tasks_base\u001b[0m]\u001b[0m\u001b[92m20:24:58 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:24:58 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:24:58 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:24:58 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:24:58 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:24:58 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:24:58 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:24:58 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:24:58 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:24:58 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:24:58 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:25:01 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:25:01 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:25:01 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:25:01 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:25:01 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:25:01 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:25:01 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:25:01 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:25:02 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:25:02 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:25:02 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:25:02 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:25:05 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:25:05 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:25:05 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:25:05 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:25:05 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:25:05 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:25:05 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:25:05 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:25:06 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:25:06 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:25:06 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:25:06 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\n",
|
||||||
"\u001b[2m2025-06-18T18:25:06.543776\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mCoroutine task started: `add_data_points`\u001b[0m [\u001b[0m\u001b[1m\u001b[34mrun_tasks_base\u001b[0m]\u001b[0m\u001b[92m20:25:07 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[2m2025-06-18T18:25:06.543776\u001b[0m [\u001b[32m\u001b[1minfo \u001b[0m] \u001b[1mCoroutine task started: `add_data_points`\u001b[0m [\u001b[0m\u001b[1m\u001b[34mrun_tasks_base\u001b[0m]\u001b[0m\u001b[92m20:25:07 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:08 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:08 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:09 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:09 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
|
|
@ -790,11 +790,11 @@
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:14 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:14 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:15 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:15 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:15 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:15 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:25:17 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:25:17 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:25:17 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:25:17 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:25:17 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:25:17 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:17 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:17 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:17 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:17 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:17 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:17 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
|
|
@ -818,11 +818,11 @@
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:20 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:20 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:20 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:20 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:20 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:20 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:25:22 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:25:22 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:25:22 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:25:22 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:25:23 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:25:23 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:23 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:23 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:23 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:23 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:23 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:23 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
|
|
@ -846,11 +846,11 @@
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:24 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:24 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:24 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:24 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:24 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:24 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:25:27 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:25:27 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:25:27 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:25:27 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:25:27 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:25:27 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:27 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:27 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:27 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:27 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:28 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:28 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
|
|
@ -874,11 +874,11 @@
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:28 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:28 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:28 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:28 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/text-embedding-3-large\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:28 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
"\u001b[1mselected model name for cost calculation: openai/text-embedding-3-large\u001b[0m\u001b[92m20:25:28 - LiteLLM:INFO\u001b[0m: utils.py:3101 - \n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\n",
|
||||||
"\u001b[1m\n",
|
"\u001b[1m\n",
|
||||||
"LiteLLM completion() model= gpt-4o-mini; provider = openai\u001b[0m\u001b[92m20:25:30 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"LiteLLM completion() model= gpt-5-mini; provider = openai\u001b[0m\u001b[92m20:25:30 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m\u001b[92m20:25:30 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\n",
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m\u001b[92m20:25:30 - LiteLLM:INFO\u001b[0m: cost_calculator.py:655 - selected model name for cost calculation: openai/gpt-5-mini-2024-07-18\n",
|
||||||
"\u001b[1mselected model name for cost calculation: openai/gpt-4o-mini-2024-07-18\u001b[0m"
|
"\u001b[1mselected model name for cost calculation: openai/gpt-5-mini-2024-07-18\u001b[0m"
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
|
|
|
||||||
1216
poetry.lock
generated
1216
poetry.lock
generated
File diff suppressed because it is too large
Load diff
|
|
@ -96,7 +96,7 @@ huggingface = ["transformers>=4.46.3,<5"]
|
||||||
ollama = ["transformers>=4.46.3,<5"]
|
ollama = ["transformers>=4.46.3,<5"]
|
||||||
mistral = ["mistral-common>=1.5.2,<2"]
|
mistral = ["mistral-common>=1.5.2,<2"]
|
||||||
anthropic = ["anthropic>=0.26.1,<0.27"]
|
anthropic = ["anthropic>=0.26.1,<0.27"]
|
||||||
deepeval = ["deepeval>=2.0.1,<3"]
|
deepeval = ["deepeval>=3.0.1,<4"]
|
||||||
posthog = ["posthog>=3.5.0,<4"]
|
posthog = ["posthog>=3.5.0,<4"]
|
||||||
falkordb = ["falkordb>=1.0.9,<2.0.0"]
|
falkordb = ["falkordb>=1.0.9,<2.0.0"]
|
||||||
groq = ["groq>=0.8.0,<1.0.0"]
|
groq = ["groq>=0.8.0,<1.0.0"]
|
||||||
|
|
|
||||||
Loading…
Add table
Reference in a new issue