Fix linter issues
This commit is contained in:
parent
60c8fd103b
commit
649fcf2ba8
9 changed files with 31 additions and 35 deletions
|
|
@ -4,7 +4,7 @@ from logging.config import fileConfig
|
|||
from sqlalchemy import pool
|
||||
from sqlalchemy.engine import Connection
|
||||
from sqlalchemy.ext.asyncio import async_engine_from_config
|
||||
|
||||
from cognee.infrastructure.databases.relational import Base
|
||||
from alembic import context
|
||||
|
||||
# this is the Alembic Config object, which provides
|
||||
|
|
@ -20,7 +20,6 @@ if config.config_file_name is not None:
|
|||
# for 'autogenerate' support
|
||||
# from myapp import mymodel
|
||||
# target_metadata = mymodel.Base.metadata
|
||||
from cognee.infrastructure.databases.relational import Base
|
||||
|
||||
target_metadata = Base.metadata
|
||||
|
||||
|
|
|
|||
|
|
@ -7,9 +7,25 @@ import sentry_sdk
|
|||
from fastapi import FastAPI, status
|
||||
from fastapi.responses import JSONResponse, Response
|
||||
from fastapi.middleware.cors import CORSMiddleware
|
||||
|
||||
from cognee.api.v1.permissions.routers import get_permissions_router
|
||||
from cognee.api.v1.settings.routers import get_settings_router
|
||||
from cognee.api.v1.datasets.routers import get_datasets_router
|
||||
from cognee.api.v1.cognify.routers import get_cognify_router
|
||||
from cognee.api.v1.search.routers import get_search_router
|
||||
from cognee.api.v1.add.routers import get_add_router
|
||||
from fastapi import Request
|
||||
from fastapi.encoders import jsonable_encoder
|
||||
from fastapi.exceptions import RequestValidationError
|
||||
from cognee.exceptions import CogneeApiError
|
||||
from traceback import format_exc
|
||||
from cognee.api.v1.users.routers import (
|
||||
get_auth_router,
|
||||
get_register_router,
|
||||
get_reset_password_router,
|
||||
get_verify_router,
|
||||
get_users_router,
|
||||
)
|
||||
from contextlib import asynccontextmanager
|
||||
|
||||
# Set up logging
|
||||
logging.basicConfig(
|
||||
|
|
@ -25,7 +41,6 @@ if os.getenv("ENV", "prod") == "prod":
|
|||
profiles_sample_rate=1.0,
|
||||
)
|
||||
|
||||
from contextlib import asynccontextmanager
|
||||
|
||||
app_environment = os.getenv("ENV", "prod")
|
||||
|
||||
|
|
@ -58,23 +73,6 @@ app.add_middleware(
|
|||
allow_headers=["*"],
|
||||
)
|
||||
|
||||
from cognee.api.v1.users.routers import (
|
||||
get_auth_router,
|
||||
get_register_router,
|
||||
get_reset_password_router,
|
||||
get_verify_router,
|
||||
get_users_router,
|
||||
)
|
||||
from cognee.api.v1.permissions.routers import get_permissions_router
|
||||
from cognee.api.v1.settings.routers import get_settings_router
|
||||
from cognee.api.v1.datasets.routers import get_datasets_router
|
||||
from cognee.api.v1.cognify.routers import get_cognify_router
|
||||
from cognee.api.v1.search.routers import get_search_router
|
||||
from cognee.api.v1.add.routers import get_add_router
|
||||
|
||||
from fastapi import Request
|
||||
from fastapi.encoders import jsonable_encoder
|
||||
from fastapi.exceptions import RequestValidationError
|
||||
|
||||
|
||||
@app.exception_handler(RequestValidationError)
|
||||
|
|
|
|||
|
|
@ -21,12 +21,12 @@ from cognee.tasks.repo_processor import (
|
|||
)
|
||||
from cognee.tasks.repo_processor.get_source_code_chunks import get_source_code_chunks
|
||||
from cognee.tasks.storage import add_data_points
|
||||
from cognee.tasks.summarization import summarize_code, summarize_text
|
||||
|
||||
monitoring = get_base_config().monitoring_tool
|
||||
if monitoring == MonitoringTool.LANGFUSE:
|
||||
from langfuse.decorators import observe
|
||||
|
||||
from cognee.tasks.summarization import summarize_code, summarize_text
|
||||
|
||||
logger = logging.getLogger("code_graph_pipeline")
|
||||
update_status_lock = asyncio.Lock()
|
||||
|
|
|
|||
|
|
@ -217,7 +217,7 @@ class FalkorDBAdapter(VectorDBInterface, GraphDBInterface):
|
|||
|
||||
async def retrieve(self, data_point_ids: list[UUID]):
|
||||
result = self.query(
|
||||
f"MATCH (node) WHERE node.id IN $node_ids RETURN node",
|
||||
"MATCH (node) WHERE node.id IN $node_ids RETURN node",
|
||||
{
|
||||
"node_ids": [str(data_point) for data_point in data_point_ids],
|
||||
},
|
||||
|
|
@ -343,7 +343,7 @@ class FalkorDBAdapter(VectorDBInterface, GraphDBInterface):
|
|||
|
||||
async def delete_data_points(self, collection_name: str, data_point_ids: list[UUID]):
|
||||
return self.query(
|
||||
f"MATCH (node) WHERE node.id IN $node_ids DETACH DELETE node",
|
||||
"MATCH (node) WHERE node.id IN $node_ids DETACH DELETE node",
|
||||
{
|
||||
"node_ids": [str(data_point) for data_point in data_point_ids],
|
||||
},
|
||||
|
|
|
|||
|
|
@ -42,7 +42,7 @@ class OpenAIAdapter(LLMInterface):
|
|||
self.endpoint = endpoint
|
||||
self.api_version = api_version
|
||||
self.streaming = streaming
|
||||
base_config = get_base_config()
|
||||
|
||||
|
||||
@observe(as_type="generation")
|
||||
async def acreate_structured_output(
|
||||
|
|
|
|||
|
|
@ -51,7 +51,7 @@ def remove_stop_words(text: str):
|
|||
|
||||
stop_words = set(stopwords.words("english"))
|
||||
text = text.split()
|
||||
text = [word for word in text if not word in stop_words]
|
||||
text = [word for word in text if word not in stop_words]
|
||||
return " ".join(text)
|
||||
|
||||
|
||||
|
|
|
|||
|
|
@ -1,8 +1,7 @@
|
|||
import logging
|
||||
|
||||
logger = logging.getLogger("task:repo_processor")
|
||||
|
||||
from .enrich_dependency_graph import enrich_dependency_graph
|
||||
from .expand_dependency_graph import expand_dependency_graph
|
||||
from .get_non_code_files import get_data_list_for_user, get_non_py_files
|
||||
from .get_repo_file_dependencies import get_repo_file_dependencies
|
||||
import logging
|
||||
|
||||
logger = logging.getLogger("task:repo_processor")
|
||||
|
|
|
|||
|
|
@ -33,14 +33,14 @@ print(dataset.goldens)
|
|||
print(dataset)
|
||||
|
||||
|
||||
import pytest
|
||||
from deepeval import assert_test
|
||||
# import pytest
|
||||
# from deepeval import assert_test
|
||||
from deepeval.metrics import AnswerRelevancyMetric
|
||||
|
||||
|
||||
answer_relevancy_metric = AnswerRelevancyMetric(threshold=0.5)
|
||||
|
||||
from deepeval import evaluate
|
||||
# from deepeval import evaluate
|
||||
|
||||
|
||||
# evaluate(dataset, [answer_relevancy_metric])
|
||||
|
|
|
|||
|
|
@ -33,7 +33,7 @@ def benchmark_function(func: Callable, *args, num_runs: int = 5) -> Dict[str, An
|
|||
start_time = time.perf_counter()
|
||||
start_cpu_time = process.cpu_times()
|
||||
|
||||
result = func(*args)
|
||||
|
||||
|
||||
end_cpu_time = process.cpu_times()
|
||||
end_time = time.perf_counter()
|
||||
|
|
@ -45,7 +45,7 @@ def benchmark_function(func: Callable, *args, num_runs: int = 5) -> Dict[str, An
|
|||
)
|
||||
current, peak = tracemalloc.get_traced_memory()
|
||||
final_memory = process.memory_info().rss
|
||||
memory_used = final_memory - initial_memory
|
||||
|
||||
|
||||
# Store results
|
||||
execution_times.append(execution_time)
|
||||
|
|
|
|||
Loading…
Add table
Reference in a new issue