Fix linter issues
This commit is contained in:
parent
649fcf2ba8
commit
6dafe73a6b
7 changed files with 19 additions and 28 deletions
|
|
@ -47,7 +47,7 @@ def get_permissions_router() -> APIRouter:
|
||||||
group_id=group.id, permission_id=permission_entity.id
|
group_id=group.id, permission_id=permission_entity.id
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
except IntegrityError as e:
|
except IntegrityError:
|
||||||
raise EntityAlreadyExistsError(message="Group permission already exists.")
|
raise EntityAlreadyExistsError(message="Group permission already exists.")
|
||||||
|
|
||||||
await db.session.commit()
|
await db.session.commit()
|
||||||
|
|
@ -70,7 +70,7 @@ def get_permissions_router() -> APIRouter:
|
||||||
# Add association directly to the association table
|
# Add association directly to the association table
|
||||||
stmt = insert(UserGroup).values(user_id=user_id, group_id=group_id)
|
stmt = insert(UserGroup).values(user_id=user_id, group_id=group_id)
|
||||||
await db.session.execute(stmt)
|
await db.session.execute(stmt)
|
||||||
except IntegrityError as e:
|
except IntegrityError:
|
||||||
raise EntityAlreadyExistsError(message="User is already part of group.")
|
raise EntityAlreadyExistsError(message="User is already part of group.")
|
||||||
|
|
||||||
await db.session.commit()
|
await db.session.commit()
|
||||||
|
|
|
||||||
|
|
@ -87,7 +87,7 @@ class LiteLLMEmbeddingEngine(EmbeddingEngine):
|
||||||
|
|
||||||
except litellm.exceptions.RateLimitError:
|
except litellm.exceptions.RateLimitError:
|
||||||
if self.retry_count >= self.MAX_RETRIES:
|
if self.retry_count >= self.MAX_RETRIES:
|
||||||
raise Exception(f"Rate limit exceeded and no more retries left.")
|
raise Exception("Rate limit exceeded and no more retries left.")
|
||||||
|
|
||||||
await exponential_backoff(self.retry_count)
|
await exponential_backoff(self.retry_count)
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -1,5 +1,6 @@
|
||||||
from typing import AsyncGenerator
|
from typing import AsyncGenerator
|
||||||
from fastapi import Depends
|
from fastapi import Depends
|
||||||
|
from contextlib import asynccontextmanager
|
||||||
from sqlalchemy.ext.asyncio import AsyncSession
|
from sqlalchemy.ext.asyncio import AsyncSession
|
||||||
from fastapi_users.db import SQLAlchemyUserDatabase
|
from fastapi_users.db import SQLAlchemyUserDatabase
|
||||||
from cognee.infrastructure.databases.relational import get_relational_engine
|
from cognee.infrastructure.databases.relational import get_relational_engine
|
||||||
|
|
@ -16,6 +17,6 @@ async def get_user_db(session: AsyncSession = Depends(get_async_session)):
|
||||||
yield SQLAlchemyUserDatabase(session, User)
|
yield SQLAlchemyUserDatabase(session, User)
|
||||||
|
|
||||||
|
|
||||||
from contextlib import asynccontextmanager
|
|
||||||
|
|
||||||
get_user_db_context = asynccontextmanager(get_user_db)
|
get_user_db_context = asynccontextmanager(get_user_db)
|
||||||
|
|
|
||||||
|
|
@ -8,17 +8,19 @@ import hashlib
|
||||||
from datetime import datetime, timezone
|
from datetime import datetime, timezone
|
||||||
import graphistry
|
import graphistry
|
||||||
import networkx as nx
|
import networkx as nx
|
||||||
import numpy as np
|
|
||||||
import pandas as pd
|
import pandas as pd
|
||||||
import matplotlib.pyplot as plt
|
import matplotlib.pyplot as plt
|
||||||
import tiktoken
|
import tiktoken
|
||||||
import nltk
|
import nltk
|
||||||
import base64
|
import base64
|
||||||
import networkx as nx
|
|
||||||
from bokeh.io import output_file, save
|
|
||||||
from bokeh.plotting import figure, from_networkx
|
from bokeh.plotting import figure, from_networkx
|
||||||
from bokeh.models import Circle, MultiLine, HoverTool, ColumnDataSource, Range1d
|
from bokeh.models import Circle, MultiLine, HoverTool, ColumnDataSource, Range1d
|
||||||
|
from bokeh.plotting import output_file, show
|
||||||
|
|
||||||
|
from bokeh.embed import file_html
|
||||||
|
from bokeh.resources import CDN
|
||||||
|
import cairosvg
|
||||||
|
import logging
|
||||||
|
|
||||||
from cognee.base_config import get_base_config
|
from cognee.base_config import get_base_config
|
||||||
from cognee.infrastructure.databases.graph import get_graph_engine
|
from cognee.infrastructure.databases.graph import get_graph_engine
|
||||||
|
|
@ -272,16 +274,7 @@ def extract_pos_tags(sentence):
|
||||||
return pos_tags
|
return pos_tags
|
||||||
|
|
||||||
|
|
||||||
import networkx as nx
|
|
||||||
from bokeh.plotting import figure, output_file, show
|
|
||||||
from bokeh.models import Circle, MultiLine, HoverTool, Range1d
|
|
||||||
from bokeh.io import output_notebook
|
|
||||||
from bokeh.embed import file_html
|
|
||||||
from bokeh.resources import CDN
|
|
||||||
from bokeh.plotting import figure, from_networkx
|
|
||||||
import base64
|
|
||||||
import cairosvg
|
|
||||||
import logging
|
|
||||||
|
|
||||||
logging.basicConfig(level=logging.INFO)
|
logging.basicConfig(level=logging.INFO)
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -49,7 +49,7 @@ async def get_graph_from_model_test():
|
||||||
for document_chunk in document_chunks:
|
for document_chunk in document_chunks:
|
||||||
document_chunk.contains.append(
|
document_chunk.contains.append(
|
||||||
Entity(
|
Entity(
|
||||||
name=f"Entity",
|
name="Entity",
|
||||||
is_type=EntityType(
|
is_type=EntityType(
|
||||||
name="Type 1",
|
name="Type 1",
|
||||||
),
|
),
|
||||||
|
|
|
||||||
|
|
@ -102,8 +102,6 @@ def test_prepare_nodes():
|
||||||
assert len(nodes_df) == 1
|
assert len(nodes_df) == 1
|
||||||
|
|
||||||
|
|
||||||
from unittest.mock import DEFAULT
|
|
||||||
|
|
||||||
|
|
||||||
def test_create_cognee_style_network_with_logo():
|
def test_create_cognee_style_network_with_logo():
|
||||||
import networkx as nx
|
import networkx as nx
|
||||||
|
|
|
||||||
|
|
@ -1,14 +1,19 @@
|
||||||
from deepeval.dataset import EvaluationDataset
|
from deepeval.dataset import EvaluationDataset
|
||||||
from pydantic import BaseModel
|
from pydantic import BaseModel
|
||||||
|
import os
|
||||||
|
|
||||||
from typing import List, Type
|
from typing import List, Type
|
||||||
from deepeval.test_case import LLMTestCase
|
from deepeval.test_case import LLMTestCase
|
||||||
import dotenv
|
import dotenv
|
||||||
|
from cognee.infrastructure.llm.get_llm_client import get_llm_client
|
||||||
|
from cognee.infrastructure.databases.vector import get_vector_engine
|
||||||
|
from cognee.base_config import get_base_config
|
||||||
|
|
||||||
|
import logging
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
dotenv.load_dotenv()
|
dotenv.load_dotenv()
|
||||||
|
|
||||||
from cognee.infrastructure.llm.get_llm_client import get_llm_client
|
|
||||||
|
|
||||||
dataset = EvaluationDataset()
|
dataset = EvaluationDataset()
|
||||||
dataset.add_test_cases_from_json_file(
|
dataset.add_test_cases_from_json_file(
|
||||||
|
|
@ -39,9 +44,6 @@ print(dataset.goldens)
|
||||||
print(dataset)
|
print(dataset)
|
||||||
|
|
||||||
|
|
||||||
import logging
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
|
|
||||||
class AnswerModel(BaseModel):
|
class AnswerModel(BaseModel):
|
||||||
|
|
@ -78,9 +80,6 @@ async def run_cognify_base_rag():
|
||||||
pass
|
pass
|
||||||
|
|
||||||
|
|
||||||
import os
|
|
||||||
from cognee.base_config import get_base_config
|
|
||||||
from cognee.infrastructure.databases.vector import get_vector_engine
|
|
||||||
|
|
||||||
|
|
||||||
async def cognify_search_base_rag(content: str, context: str):
|
async def cognify_search_base_rag(content: str, context: str):
|
||||||
|
|
|
||||||
Loading…
Add table
Reference in a new issue