From e68b19ec835da2009ad3aa763c6e6521b4a1e3c7 Mon Sep 17 00:00:00 2001 From: hajdul88 <52442977+hajdul88@users.noreply.github.com> Date: Wed, 2 Jul 2025 18:56:11 +0200 Subject: [PATCH] changes modal entry + cognify batching --- 1gb_nonparallel_cognee.py | 24 ++++++++++++++---------- cognee/api/v1/cognify/cognify.py | 6 +++--- 2 files changed, 17 insertions(+), 13 deletions(-) diff --git a/1gb_nonparallel_cognee.py b/1gb_nonparallel_cognee.py index 7e5aad23f..0a9d83bf4 100644 --- a/1gb_nonparallel_cognee.py +++ b/1gb_nonparallel_cognee.py @@ -12,20 +12,24 @@ image = ( modal.Image.from_dockerfile(path="Dockerfile_modal", force_build=False) .copy_local_file("pyproject.toml", "pyproject.toml") .copy_local_file("poetry.lock", "poetry.lock") - .env( - { - "ENV": os.getenv("ENV"), - "LLM_API_KEY": os.getenv("LLM_API_KEY"), - "OPENAI_API_KEY": os.getenv("OPENAI_API_KEY"), - "AWS_ACCESS_KEY_ID": os.getenv("AWS_ACCESS_KEY_ID"), - "AWS_SECRET_ACCESS_KEY": os.getenv("AWS_SECRET_ACCESS_KEY"), - } + .pip_install( + "protobuf", + "h2", + "deepeval", + "gdown", + "plotly", + "psycopg2-binary==2.9.10", + "asyncpg==0.30.0", ) - .pip_install("protobuf", "h2", "deepeval", "gdown", "plotly") ) -@app.function(image=image, max_containers=1, timeout=86400, min_containers=1) +@app.function( + image=image, + max_containers=1, + timeout=86400, + secrets=[modal.Secret.from_name("1gb_nonparallel_cognee")], +) async def run_cognee_1gb(): await cognee.prune.prune_data() await cognee.prune.prune_system(metadata=True) diff --git a/cognee/api/v1/cognify/cognify.py b/cognee/api/v1/cognify/cognify.py index a3cf645d3..8528fd608 100644 --- a/cognee/api/v1/cognify/cognify.py +++ b/cognee/api/v1/cognify/cognify.py @@ -151,13 +151,13 @@ async def get_default_tasks( # TODO: Find out a better way to do this (Boris's extract_graph_from_data, graph_model=graph_model, ontology_adapter=OntologyResolver(ontology_file=ontology_file_path), - task_config={"batch_size": 10}, + task_config={"batch_size": 50}, ), # Generate knowledge graphs from the document chunks. Task( summarize_text, - task_config={"batch_size": 10}, + task_config={"batch_size": 50}, ), - Task(add_data_points, task_config={"batch_size": 10}), + Task(add_data_points, task_config={"batch_size": 50}), ] return default_tasks