Update litellm and openai version, bump pydantic
This commit is contained in:
parent
cc77d844b6
commit
287427a240
11 changed files with 441 additions and 436 deletions
|
|
@ -146,7 +146,7 @@ class DatasetDTO(OutDTO):
|
|||
id: UUID
|
||||
name: str
|
||||
created_at: datetime
|
||||
updated_at: Optional[datetime]
|
||||
updated_at: Optional[datetime] = None
|
||||
owner_id: UUID
|
||||
|
||||
@app.get("/api/v1/datasets", response_model = list[DatasetDTO])
|
||||
|
|
@ -200,7 +200,7 @@ class DataDTO(OutDTO):
|
|||
id: UUID
|
||||
name: str
|
||||
created_at: datetime
|
||||
updated_at: Optional[datetime]
|
||||
updated_at: Optional[datetime] = None
|
||||
extension: str
|
||||
mime_type: str
|
||||
raw_data_location: str
|
||||
|
|
|
|||
|
|
@ -5,7 +5,7 @@ from .models.Task import Task
|
|||
|
||||
class PipelineConfig(BaseModel):
|
||||
batch_count: int = 10
|
||||
description: Optional[str]
|
||||
description: Optional[str] = None
|
||||
|
||||
class Pipeline():
|
||||
id: UUID = uuid4()
|
||||
|
|
|
|||
|
|
@ -1,8 +1,8 @@
|
|||
from typing import Any, Callable, Generator
|
||||
from typing import Any, Callable, Generator, List
|
||||
import asyncio
|
||||
from ..tasks.Task import Task
|
||||
|
||||
def run_tasks_parallel(tasks: [Task]) -> Callable[[Any], Generator[Any, Any, Any]]:
|
||||
def run_tasks_parallel(tasks: List[Task]) -> Callable[[Any], Generator[Any, Any, Any]]:
|
||||
async def parallel_run(*args, **kwargs):
|
||||
parallel_tasks = [asyncio.create_task(task.run(*args, **kwargs)) for task in tasks]
|
||||
|
||||
|
|
|
|||
|
|
@ -6,15 +6,15 @@ class BaseClass(BaseModel):
|
|||
name: str
|
||||
type: Literal["Class"] = "Class"
|
||||
description: str
|
||||
constructor_parameters: Optional[List[str]]
|
||||
constructor_parameters: Optional[List[str]] = None
|
||||
|
||||
class Class(BaseModel):
|
||||
id: str
|
||||
name: str
|
||||
type: Literal["Class"] = "Class"
|
||||
description: str
|
||||
constructor_parameters: Optional[List[str]]
|
||||
from_class: Optional[BaseClass]
|
||||
constructor_parameters: Optional[List[str]] = None
|
||||
from_class: Optional[BaseClass] = None
|
||||
|
||||
class ClassInstance(BaseModel):
|
||||
id: str
|
||||
|
|
@ -28,7 +28,7 @@ class Function(BaseModel):
|
|||
name: str
|
||||
type: Literal["Function"] = "Function"
|
||||
description: str
|
||||
parameters: Optional[List[str]]
|
||||
parameters: Optional[List[str]] = None
|
||||
return_type: str
|
||||
is_static: Optional[bool] = False
|
||||
|
||||
|
|
@ -38,7 +38,7 @@ class Variable(BaseModel):
|
|||
type: Literal["Variable"] = "Variable"
|
||||
description: str
|
||||
is_static: Optional[bool] = False
|
||||
default_value: Optional[str]
|
||||
default_value: Optional[str] = None
|
||||
|
||||
class Operator(BaseModel):
|
||||
id: str
|
||||
|
|
|
|||
|
|
@ -49,7 +49,7 @@ async def main():
|
|||
|
||||
search_results = await cognee.search(SearchType.SUMMARIES, query = random_node_name)
|
||||
assert len(search_results) != 0, "Query related summaries don't exist."
|
||||
print("\n\Extracted summaries are:\n")
|
||||
print("\nExtracted summaries are:\n")
|
||||
for result in search_results:
|
||||
print(f"{result}\n")
|
||||
|
||||
|
|
|
|||
|
|
@ -53,7 +53,7 @@ async def main():
|
|||
|
||||
search_results = await cognee.search(SearchType.SUMMARIES, query = random_node_name)
|
||||
assert len(search_results) != 0, "Query related summaries don't exist."
|
||||
print("\n\Extracted summaries are:\n")
|
||||
print("\nExtracted summaries are:\n")
|
||||
for result in search_results:
|
||||
print(f"{result}\n")
|
||||
|
||||
|
|
|
|||
|
|
@ -54,7 +54,7 @@ async def main():
|
|||
|
||||
search_results = await cognee.search(SearchType.SUMMARIES, query = random_node_name)
|
||||
assert len(search_results) != 0, "Query related summaries don't exist."
|
||||
print("\n\Extracted summaries are:\n")
|
||||
print("\nExtracted summaries are:\n")
|
||||
for result in search_results:
|
||||
print(f"{result}\n")
|
||||
|
||||
|
|
|
|||
|
|
@ -52,7 +52,7 @@ async def main():
|
|||
|
||||
search_results = await cognee.search(SearchType.SUMMARIES, query = random_node_name)
|
||||
assert len(search_results) != 0, "Query related summaries don't exist."
|
||||
print("\n\Extracted summaries are:\n")
|
||||
print("\nExtracted summaries are:\n")
|
||||
for result in search_results:
|
||||
print(f"{result}\n")
|
||||
|
||||
|
|
|
|||
0
log.txt
Normal file
0
log.txt
Normal file
845
poetry.lock
generated
845
poetry.lock
generated
File diff suppressed because it is too large
Load diff
|
|
@ -19,7 +19,7 @@ classifiers = [
|
|||
|
||||
[tool.poetry.dependencies]
|
||||
python = ">=3.9.0,<3.12"
|
||||
openai = "1.27.0"
|
||||
openai = "1.52.0"
|
||||
pydantic = "2.8.2"
|
||||
python-dotenv = "1.0.1"
|
||||
fastapi = "^0.109.2"
|
||||
|
|
@ -55,7 +55,7 @@ structlog = "^24.1.0"
|
|||
tiktoken = "0.7.0"
|
||||
posthog = "^3.5.0"
|
||||
lancedb = "0.8.0"
|
||||
litellm = "1.38.10"
|
||||
litellm = "1.51.3"
|
||||
groq = "0.8.0"
|
||||
tantivy = "^0.22.0"
|
||||
tokenizers ="0.15.2"
|
||||
|
|
|
|||
Loading…
Add table
Reference in a new issue