Update litellm and openai version, bump pydantic
This commit is contained in:
parent
cc77d844b6
commit
287427a240
11 changed files with 441 additions and 436 deletions
|
|
@ -146,7 +146,7 @@ class DatasetDTO(OutDTO):
|
||||||
id: UUID
|
id: UUID
|
||||||
name: str
|
name: str
|
||||||
created_at: datetime
|
created_at: datetime
|
||||||
updated_at: Optional[datetime]
|
updated_at: Optional[datetime] = None
|
||||||
owner_id: UUID
|
owner_id: UUID
|
||||||
|
|
||||||
@app.get("/api/v1/datasets", response_model = list[DatasetDTO])
|
@app.get("/api/v1/datasets", response_model = list[DatasetDTO])
|
||||||
|
|
@ -200,7 +200,7 @@ class DataDTO(OutDTO):
|
||||||
id: UUID
|
id: UUID
|
||||||
name: str
|
name: str
|
||||||
created_at: datetime
|
created_at: datetime
|
||||||
updated_at: Optional[datetime]
|
updated_at: Optional[datetime] = None
|
||||||
extension: str
|
extension: str
|
||||||
mime_type: str
|
mime_type: str
|
||||||
raw_data_location: str
|
raw_data_location: str
|
||||||
|
|
|
||||||
|
|
@ -5,7 +5,7 @@ from .models.Task import Task
|
||||||
|
|
||||||
class PipelineConfig(BaseModel):
|
class PipelineConfig(BaseModel):
|
||||||
batch_count: int = 10
|
batch_count: int = 10
|
||||||
description: Optional[str]
|
description: Optional[str] = None
|
||||||
|
|
||||||
class Pipeline():
|
class Pipeline():
|
||||||
id: UUID = uuid4()
|
id: UUID = uuid4()
|
||||||
|
|
|
||||||
|
|
@ -1,8 +1,8 @@
|
||||||
from typing import Any, Callable, Generator
|
from typing import Any, Callable, Generator, List
|
||||||
import asyncio
|
import asyncio
|
||||||
from ..tasks.Task import Task
|
from ..tasks.Task import Task
|
||||||
|
|
||||||
def run_tasks_parallel(tasks: [Task]) -> Callable[[Any], Generator[Any, Any, Any]]:
|
def run_tasks_parallel(tasks: List[Task]) -> Callable[[Any], Generator[Any, Any, Any]]:
|
||||||
async def parallel_run(*args, **kwargs):
|
async def parallel_run(*args, **kwargs):
|
||||||
parallel_tasks = [asyncio.create_task(task.run(*args, **kwargs)) for task in tasks]
|
parallel_tasks = [asyncio.create_task(task.run(*args, **kwargs)) for task in tasks]
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -6,15 +6,15 @@ class BaseClass(BaseModel):
|
||||||
name: str
|
name: str
|
||||||
type: Literal["Class"] = "Class"
|
type: Literal["Class"] = "Class"
|
||||||
description: str
|
description: str
|
||||||
constructor_parameters: Optional[List[str]]
|
constructor_parameters: Optional[List[str]] = None
|
||||||
|
|
||||||
class Class(BaseModel):
|
class Class(BaseModel):
|
||||||
id: str
|
id: str
|
||||||
name: str
|
name: str
|
||||||
type: Literal["Class"] = "Class"
|
type: Literal["Class"] = "Class"
|
||||||
description: str
|
description: str
|
||||||
constructor_parameters: Optional[List[str]]
|
constructor_parameters: Optional[List[str]] = None
|
||||||
from_class: Optional[BaseClass]
|
from_class: Optional[BaseClass] = None
|
||||||
|
|
||||||
class ClassInstance(BaseModel):
|
class ClassInstance(BaseModel):
|
||||||
id: str
|
id: str
|
||||||
|
|
@ -28,7 +28,7 @@ class Function(BaseModel):
|
||||||
name: str
|
name: str
|
||||||
type: Literal["Function"] = "Function"
|
type: Literal["Function"] = "Function"
|
||||||
description: str
|
description: str
|
||||||
parameters: Optional[List[str]]
|
parameters: Optional[List[str]] = None
|
||||||
return_type: str
|
return_type: str
|
||||||
is_static: Optional[bool] = False
|
is_static: Optional[bool] = False
|
||||||
|
|
||||||
|
|
@ -38,7 +38,7 @@ class Variable(BaseModel):
|
||||||
type: Literal["Variable"] = "Variable"
|
type: Literal["Variable"] = "Variable"
|
||||||
description: str
|
description: str
|
||||||
is_static: Optional[bool] = False
|
is_static: Optional[bool] = False
|
||||||
default_value: Optional[str]
|
default_value: Optional[str] = None
|
||||||
|
|
||||||
class Operator(BaseModel):
|
class Operator(BaseModel):
|
||||||
id: str
|
id: str
|
||||||
|
|
|
||||||
|
|
@ -49,7 +49,7 @@ async def main():
|
||||||
|
|
||||||
search_results = await cognee.search(SearchType.SUMMARIES, query = random_node_name)
|
search_results = await cognee.search(SearchType.SUMMARIES, query = random_node_name)
|
||||||
assert len(search_results) != 0, "Query related summaries don't exist."
|
assert len(search_results) != 0, "Query related summaries don't exist."
|
||||||
print("\n\Extracted summaries are:\n")
|
print("\nExtracted summaries are:\n")
|
||||||
for result in search_results:
|
for result in search_results:
|
||||||
print(f"{result}\n")
|
print(f"{result}\n")
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -53,7 +53,7 @@ async def main():
|
||||||
|
|
||||||
search_results = await cognee.search(SearchType.SUMMARIES, query = random_node_name)
|
search_results = await cognee.search(SearchType.SUMMARIES, query = random_node_name)
|
||||||
assert len(search_results) != 0, "Query related summaries don't exist."
|
assert len(search_results) != 0, "Query related summaries don't exist."
|
||||||
print("\n\Extracted summaries are:\n")
|
print("\nExtracted summaries are:\n")
|
||||||
for result in search_results:
|
for result in search_results:
|
||||||
print(f"{result}\n")
|
print(f"{result}\n")
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -54,7 +54,7 @@ async def main():
|
||||||
|
|
||||||
search_results = await cognee.search(SearchType.SUMMARIES, query = random_node_name)
|
search_results = await cognee.search(SearchType.SUMMARIES, query = random_node_name)
|
||||||
assert len(search_results) != 0, "Query related summaries don't exist."
|
assert len(search_results) != 0, "Query related summaries don't exist."
|
||||||
print("\n\Extracted summaries are:\n")
|
print("\nExtracted summaries are:\n")
|
||||||
for result in search_results:
|
for result in search_results:
|
||||||
print(f"{result}\n")
|
print(f"{result}\n")
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -52,7 +52,7 @@ async def main():
|
||||||
|
|
||||||
search_results = await cognee.search(SearchType.SUMMARIES, query = random_node_name)
|
search_results = await cognee.search(SearchType.SUMMARIES, query = random_node_name)
|
||||||
assert len(search_results) != 0, "Query related summaries don't exist."
|
assert len(search_results) != 0, "Query related summaries don't exist."
|
||||||
print("\n\Extracted summaries are:\n")
|
print("\nExtracted summaries are:\n")
|
||||||
for result in search_results:
|
for result in search_results:
|
||||||
print(f"{result}\n")
|
print(f"{result}\n")
|
||||||
|
|
||||||
|
|
|
||||||
0
log.txt
Normal file
0
log.txt
Normal file
845
poetry.lock
generated
845
poetry.lock
generated
File diff suppressed because it is too large
Load diff
|
|
@ -19,7 +19,7 @@ classifiers = [
|
||||||
|
|
||||||
[tool.poetry.dependencies]
|
[tool.poetry.dependencies]
|
||||||
python = ">=3.9.0,<3.12"
|
python = ">=3.9.0,<3.12"
|
||||||
openai = "1.27.0"
|
openai = "1.52.0"
|
||||||
pydantic = "2.8.2"
|
pydantic = "2.8.2"
|
||||||
python-dotenv = "1.0.1"
|
python-dotenv = "1.0.1"
|
||||||
fastapi = "^0.109.2"
|
fastapi = "^0.109.2"
|
||||||
|
|
@ -55,7 +55,7 @@ structlog = "^24.1.0"
|
||||||
tiktoken = "0.7.0"
|
tiktoken = "0.7.0"
|
||||||
posthog = "^3.5.0"
|
posthog = "^3.5.0"
|
||||||
lancedb = "0.8.0"
|
lancedb = "0.8.0"
|
||||||
litellm = "1.38.10"
|
litellm = "1.51.3"
|
||||||
groq = "0.8.0"
|
groq = "0.8.0"
|
||||||
tantivy = "^0.22.0"
|
tantivy = "^0.22.0"
|
||||||
tokenizers ="0.15.2"
|
tokenizers ="0.15.2"
|
||||||
|
|
|
||||||
Loading…
Add table
Reference in a new issue