chore: rename package in files

This commit is contained in:
Boris Arzentar 2024-03-13 16:27:07 +01:00
parent 40e760f517
commit d5391f903c
66 changed files with 6022 additions and 6022 deletions

View file

@ -7,7 +7,7 @@ POSTGRES_PASSWORD = bla
POSTGRES_DB = bubu
POSTGRES_HOST = localhost
POSTGRES_HOST_DOCKER = postgres
COG_ARCH_DIR = cognitive_architecture
COG_ARCH_DIR = cognee
GRAPH_DB_URL =
GRAPH_DB_PW =
GRAPH_DB_USER =

2
.gitignore vendored
View file

@ -164,4 +164,4 @@ cython_debug/
.vscode/
database/data/
cognitive_architecture/data/
cognee/data/

File diff suppressed because one or more lines are too long

View file

@ -43,7 +43,7 @@ WORKDIR /app
# Set the PYTHONPATH environment variable to include the /app directory
ENV PYTHONPATH=/app
COPY cognitive_architecture/ /app/cognitive_architecture
COPY cognee/ /app/cognee
COPY main.py /app
COPY api.py /app

20
api.py
View file

@ -13,7 +13,7 @@ logging.basicConfig(
logger = logging.getLogger(__name__)
from cognitive_architecture.config import Config
from cognee.config import Config
config = Config()
config.load()
@ -23,9 +23,9 @@ from fastapi import FastAPI, BackgroundTasks, HTTPException
from fastapi.responses import JSONResponse
from pydantic import BaseModel
from cognitive_architecture.database.relationaldb.database import AsyncSessionLocal
from cognitive_architecture.database.relationaldb.database_crud import session_scope
from cognitive_architecture.vectorstore_manager import Memory
from cognee.database.relationaldb.database import AsyncSessionLocal
from cognee.database.relationaldb.database_crud import session_scope
from cognee.vectorstore_manager import Memory
from main import add_documents_to_graph_db, user_context_enrichment
OPENAI_API_KEY = os.getenv("OPENAI_API_KEY", "")
@ -57,7 +57,7 @@ def health_check():
class Payload(BaseModel):
payload: Dict[str, Any]
from cognitive_architecture.api.v1.memory.create_memory import MemoryType
from cognee.api.v1.memory.create_memory import MemoryType
class CreateMemoryPayload(BaseModel):
user_id: str
@ -66,7 +66,7 @@ class CreateMemoryPayload(BaseModel):
@app.post("/create-memory", response_model=dict)
async def create_memory(payload: CreateMemoryPayload):
from cognitive_architecture.api.v1.memory.create_memory import create_memory as create_memory_v1, MemoryException
from cognee.api.v1.memory.create_memory import create_memory as create_memory_v1, MemoryException
try:
await create_memory_v1(
@ -93,7 +93,7 @@ class RememberPayload(BaseModel):
@app.post("/remember", response_model=dict)
async def remember(payload: RememberPayload):
from cognitive_architecture.api.v1.memory.remember import remember as remember_v1
from cognee.api.v1.memory.remember import remember as remember_v1
await remember_v1(
payload.user_id,
@ -264,7 +264,7 @@ async def user_query_classfier(payload: Payload):
# Execute the query - replace this with the actual execution method
async with session_scope(session=AsyncSessionLocal()) as session:
from cognitive_architecture.classifiers.classify_user_input import (
from cognee.classifiers.classify_user_input import (
classify_user_query,
)
@ -292,7 +292,7 @@ async def drop_db(payload: Payload):
else:
pass
from cognitive_architecture.database.create_database import (
from cognee.database.create_database import (
drop_database,
create_admin_engine,
)
@ -310,7 +310,7 @@ async def drop_db(payload: Payload):
else:
pass
from cognitive_architecture.database.create_database import (
from cognee.database.create_database import (
create_database,
create_admin_engine,
)

File diff suppressed because one or more lines are too long

View file

@ -1,94 +1,94 @@
{
"cells": [
{
"cell_type": "code",
"execution_count": null,
"id": "c4d5a399",
"metadata": {},
"outputs": [],
"source": [
"from os import listdir, path\n",
"from uuid import uuid5, UUID\n",
"from cognitive_architecture import add\n",
"\n",
"data_path = path.abspath(\".data\")\n",
"\n",
"results = await add(data_path, \"izmene\")\n",
"for result in results:\n",
" print(result)\n"
]
"cells": [
{
"cell_type": "code",
"execution_count": null,
"id": "c4d5a399",
"metadata": {},
"outputs": [],
"source": [
"from os import listdir, path\n",
"from uuid import uuid5, UUID\n",
"from cognee import add\n",
"\n",
"data_path = path.abspath(\".data\")\n",
"\n",
"results = await add(data_path, \"izmene\")\n",
"for result in results:\n",
" print(result)\n"
]
},
{
"cell_type": "code",
"execution_count": null,
"id": "47edce91",
"metadata": {},
"outputs": [],
"source": [
"import duckdb\n",
"from cognee.root_dir import get_absolute_path\n",
"\n",
"dataset_name = \"pdf_files\"\n",
"\n",
"db_path = get_absolute_path(\"./data/cognee\")\n",
"db_location = db_path + \"/cognee.duckdb\"\n",
"print(db_location)\n",
"\n",
"db = duckdb.connect(db_location)\n",
"\n",
"tables = db.sql(\"SELECT DISTINCT schema_name FROM duckdb_tables();\").df()\n",
"print(list(filter(lambda table_name: table_name.endswith('staging') is False, tables.to_dict()[\"schema_name\"].values())))\n",
"\n",
"# izmene = db.sql(f\"SELECT * FROM izmene.file_metadata;\")\n",
"\n",
"# print(izmene)\n",
"\n",
"# pravilnik = db.sql(f\"SELECT * FROM pravilnik.file_metadata;\")\n",
"\n",
"# print(pravilnik)\n",
"\n",
"\n"
]
},
{
"cell_type": "code",
"execution_count": null,
"id": "607bf624",
"metadata": {},
"outputs": [],
"source": [
"from os import path, listdir\n",
"from cognee import cognify, list_datasets\n",
"from cognee.utils import render_graph\n",
"\n",
"print(list_datasets())\n",
"\n",
"graph = await cognify(\"izmene\")\n",
"\n",
"graph_url = await render_graph(graph, graph_type = \"networkx\")\n",
"print(graph_url)\n"
]
}
],
"metadata": {
"kernelspec": {
"display_name": "Python 3 (ipykernel)",
"language": "python",
"name": "python3"
},
"language_info": {
"codemirror_mode": {
"name": "ipython",
"version": 3
},
"file_extension": ".py",
"mimetype": "text/x-python",
"name": "python",
"nbconvert_exporter": "python",
"pygments_lexer": "ipython3",
"version": "3.10.13"
}
},
{
"cell_type": "code",
"execution_count": null,
"id": "47edce91",
"metadata": {},
"outputs": [],
"source": [
"import duckdb\n",
"from cognitive_architecture.root_dir import get_absolute_path\n",
"\n",
"dataset_name = \"pdf_files\"\n",
"\n",
"db_path = get_absolute_path(\"./data/cognee\")\n",
"db_location = db_path + \"/cognee.duckdb\"\n",
"print(db_location)\n",
"\n",
"db = duckdb.connect(db_location)\n",
"\n",
"tables = db.sql(\"SELECT DISTINCT schema_name FROM duckdb_tables();\").df()\n",
"print(list(filter(lambda table_name: table_name.endswith('staging') is False, tables.to_dict()[\"schema_name\"].values())))\n",
"\n",
"# izmene = db.sql(f\"SELECT * FROM izmene.file_metadata;\")\n",
"\n",
"# print(izmene)\n",
"\n",
"# pravilnik = db.sql(f\"SELECT * FROM pravilnik.file_metadata;\")\n",
"\n",
"# print(pravilnik)\n",
"\n",
"\n"
]
},
{
"cell_type": "code",
"execution_count": null,
"id": "607bf624",
"metadata": {},
"outputs": [],
"source": [
"from os import path, listdir\n",
"from cognitive_architecture import cognify, list_datasets\n",
"from cognitive_architecture.utils import render_graph\n",
"\n",
"print(list_datasets())\n",
"\n",
"graph = await cognify(\"izmene\")\n",
"\n",
"graph_url = await render_graph(graph, graph_type = \"networkx\")\n",
"print(graph_url)\n"
]
}
],
"metadata": {
"kernelspec": {
"display_name": "Python 3 (ipykernel)",
"language": "python",
"name": "python3"
},
"language_info": {
"codemirror_mode": {
"name": "ipython",
"version": 3
},
"file_extension": ".py",
"mimetype": "text/x-python",
"name": "python",
"nbconvert_exporter": "python",
"pygments_lexer": "ipython3",
"version": "3.10.13"
}
},
"nbformat": 4,
"nbformat_minor": 5
}
"nbformat": 4,
"nbformat_minor": 5
}

View file

@ -4,10 +4,10 @@ import asyncio
import dlt
import duckdb
from unstructured.cleaners.core import clean
from cognitive_architecture.root_dir import get_absolute_path
import cognitive_architecture.modules.ingestion as ingestion
from cognitive_architecture.infrastructure.files import get_file_metadata
from cognitive_architecture.infrastructure.files.storage import LocalStorage
from cognee.root_dir import get_absolute_path
import cognee.modules.ingestion as ingestion
from cognee.infrastructure.files import get_file_metadata
from cognee.infrastructure.files.storage import LocalStorage
async def add(file_paths: Union[str, List[str]], dataset_name: str = None):
if isinstance(file_paths, str):

View file

@ -1,8 +1,8 @@
import asyncio
from uuid import UUID, uuid4
from typing import Union, BinaryIO, List
import cognitive_architecture.modules.ingestion as ingestion
from cognitive_architecture.infrastructure import infrastructure_config
import cognee.modules.ingestion as ingestion
from cognee.infrastructure import infrastructure_config
class DatasetException(Exception):
message: str

View file

@ -1,6 +1,6 @@
from typing import List
from enum import Enum
from cognitive_architecture.modules.users.memory import create_information_points, is_existing_memory
from cognee.modules.users.memory import create_information_points, is_existing_memory
class MemoryType(Enum):
GRAPH = "GRAPH"

View file

@ -6,27 +6,27 @@ import instructor
from openai import OpenAI
from unstructured.cleaners.core import clean
from unstructured.partition.pdf import partition_pdf
from cognitive_architecture.infrastructure.databases.vector.qdrant.QDrantAdapter import CollectionConfig
from cognitive_architecture.infrastructure.llm.get_llm_client import get_llm_client
from cognitive_architecture.modules.cognify.graph.add_classification_nodes import add_classification_nodes
from cognitive_architecture.modules.cognify.graph.add_node_connections import add_node_connection, graph_ready_output, \
from cognee.infrastructure.databases.vector.qdrant.QDrantAdapter import CollectionConfig
from cognee.infrastructure.llm.get_llm_client import get_llm_client
from cognee.modules.cognify.graph.add_classification_nodes import add_classification_nodes
from cognee.modules.cognify.graph.add_node_connections import add_node_connection, graph_ready_output, \
connect_nodes_in_graph, extract_node_descriptions
from cognitive_architecture.modules.cognify.graph.add_propositions import append_to_graph
from cognitive_architecture.modules.cognify.llm.add_node_connection_embeddings import process_items
from cognitive_architecture.modules.cognify.vector.batch_search import adapted_qdrant_batch_search
from cognitive_architecture.modules.cognify.vector.add_propositions import add_propositions
from cognee.modules.cognify.graph.add_propositions import append_to_graph
from cognee.modules.cognify.llm.add_node_connection_embeddings import process_items
from cognee.modules.cognify.vector.batch_search import adapted_qdrant_batch_search
from cognee.modules.cognify.vector.add_propositions import add_propositions
from cognitive_architecture.config import Config
from cognitive_architecture.modules.cognify.llm.classify_content import classify_into_categories
from cognitive_architecture.modules.cognify.llm.content_to_cog_layers import content_to_cog_layers
from cognitive_architecture.modules.cognify.llm.generate_graph import generate_graph
from cognitive_architecture.shared.data_models import DefaultContentPrediction, KnowledgeGraph, DefaultCognitiveLayer
from cognitive_architecture.infrastructure.databases.graph.get_graph_client import get_graph_client
from cognitive_architecture.shared.data_models import GraphDBType
from cognitive_architecture.infrastructure.databases.vector.get_vector_database import get_vector_database
from cognitive_architecture.infrastructure.databases.relational import DuckDBAdapter
from cognitive_architecture.modules.cognify.graph.add_document_node import add_document_node
from cognitive_architecture.modules.cognify.graph.initialize_graph import initialize_graph
from cognee.config import Config
from cognee.modules.cognify.llm.classify_content import classify_into_categories
from cognee.modules.cognify.llm.content_to_cog_layers import content_to_cog_layers
from cognee.modules.cognify.llm.generate_graph import generate_graph
from cognee.shared.data_models import DefaultContentPrediction, KnowledgeGraph, DefaultCognitiveLayer
from cognee.infrastructure.databases.graph.get_graph_client import get_graph_client
from cognee.shared.data_models import GraphDBType
from cognee.infrastructure.databases.vector.get_vector_database import get_vector_database
from cognee.infrastructure.databases.relational import DuckDBAdapter
from cognee.modules.cognify.graph.add_document_node import add_document_node
from cognee.modules.cognify.graph.initialize_graph import initialize_graph
config = Config()
config.load()

View file

@ -1,5 +1,5 @@
from cognitive_architecture.infrastructure.databases.relational import DuckDBAdapter
from cognee.infrastructure.databases.relational import DuckDBAdapter
def list_datasets():
db = DuckDBAdapter()

View file

@ -2,12 +2,12 @@
from enum import Enum, auto
from typing import Dict, Any, Callable, List
from cognitive_architecture.infrastructure.databases.graph.get_graph_client import get_graph_client
from cognitive_architecture.modules.search.graph.search_adjacent import search_adjacent
from cognitive_architecture.modules.search.vector.search_similarity import search_similarity
from cognitive_architecture.modules.search.graph.search_categories import search_categories
from cognitive_architecture.modules.search.graph.search_neighbour import search_neighbour
from cognitive_architecture.shared.data_models import GraphDBType
from cognee.infrastructure.databases.graph.get_graph_client import get_graph_client
from cognee.modules.search.graph.search_adjacent import search_adjacent
from cognee.modules.search.vector.search_similarity import search_similarity
from cognee.modules.search.graph.search_categories import search_categories
from cognee.modules.search.graph.search_neighbour import search_neighbour
from cognee.shared.data_models import GraphDBType
class SearchType(Enum):

View file

@ -4,8 +4,8 @@ import os
from contextlib import asynccontextmanager
from sqlalchemy import text
from sqlalchemy.ext.asyncio import create_async_engine
from cognitive_architecture.config import Config
from cognitive_architecture.database.relationaldb.database import Base, get_sqlalchemy_database_url
from cognee.config import Config
from cognee.database.relationaldb.database import Base, get_sqlalchemy_database_url
globalConfig = Config()

View file

@ -21,7 +21,7 @@ from ...utils import (
create_edge_variable_mapping,
create_node_variable_mapping,
)
from cognitive_architecture.infrastructure.llm.openai.queries import generate_summary, generate_graph
from cognee.infrastructure.llm.openai.queries import generate_summary, generate_graph
import logging
from neo4j import AsyncGraphDatabase
from contextlib import asynccontextmanager

View file

@ -1,6 +1,6 @@
import pickle
from pathlib import Path
from cognitive_architecture.config import Config
from cognee.config import Config
import networkx as nx
config = Config()
config = config.load()

View file

@ -3,7 +3,7 @@ from pathlib import Path
# from contextlib import asynccontextmanager
from sqlalchemy.ext.asyncio import create_async_engine, async_sessionmaker, AsyncSession
from sqlalchemy.orm import declarative_base
from cognitive_architecture.config import Config
from cognee.config import Config
globalConfig = Config()

View file

@ -16,12 +16,12 @@ from langchain.retrievers import WeaviateHybridSearchRetriever
from weaviate.gql.get import HybridFusion
from cognitive_architecture.database.relationaldb.models.sessions import Session
from cognitive_architecture.database.relationaldb.models.metadatas import MetaDatas
from cognitive_architecture.database.relationaldb.models.operation import Operation
from cognitive_architecture.database.relationaldb.models.docs import DocsModel
from cognee.database.relationaldb.models.sessions import Session
from cognee.database.relationaldb.models.metadatas import MetaDatas
from cognee.database.relationaldb.models.operation import Operation
from cognee.database.relationaldb.models.docs import DocsModel
from sqlalchemy.orm import sessionmaker
from cognitive_architecture.database.relationaldb.database import engine
from cognee.database.relationaldb.database import engine
from typing import Optional
import time
@ -31,7 +31,7 @@ tracemalloc.start()
from datetime import datetime
from langchain.embeddings.openai import OpenAIEmbeddings
from cognitive_architecture.database.vectordb.vectordb import (
from cognee.database.vectordb.vectordb import (
PineconeVectorDB,
WeaviateVectorDB,
LanceDB,
@ -41,7 +41,7 @@ import uuid
import weaviate
from marshmallow import Schema, fields
import json
from cognitive_architecture.database.vectordb.vector_db_type import VectorDBType
from cognee.database.vectordb.vector_db_type import VectorDBType
OPENAI_API_KEY = os.getenv("OPENAI_API_KEY", "")

View file

@ -2,7 +2,7 @@
import re
import logging
from cognitive_architecture.database.vectordb.chunkers.chunk_strategy import ChunkStrategy
from cognee.database.vectordb.chunkers.chunk_strategy import ChunkStrategy
from langchain.text_splitter import RecursiveCharacterTextSplitter

View file

@ -3,8 +3,8 @@ import fitz
import os
import sys
from cognitive_architecture.database.vectordb.chunkers.chunkers import chunk_data
from cognitive_architecture.shared.language_processing import (
from cognee.database.vectordb.chunkers.chunkers import chunk_data
from cognee.shared.language_processing import (
translate_text,
detect_language,
)
@ -148,7 +148,7 @@ async def _document_loader(observation: str, loader_settings: dict):
# file_content += page.get_text()
# pages = chunk_data(chunk_strategy=loader_strategy, source_data=file_content, chunk_size=chunk_size,
# chunk_overlap=chunk_overlap)
# from cognitive_architecture.shared.language_processing import translate_text,detect_language
# from cognee.shared.language_processing import translate_text,detect_language
#
# if detect_language(pages) != "en":
# logging.info("Current Directory 3")
@ -169,7 +169,7 @@ async def _document_loader(observation: str, loader_settings: dict):
# pages = chunk_data(chunk_strategy=loader_strategy, source_data=file_content, chunk_size=chunk_size,
# chunk_overlap=chunk_overlap)
#
# from cognitive_architecture.shared.language_processing import translate_text, detect_language
# from cognee.shared.language_processing import translate_text, detect_language
#
# if detect_language(pages) != "en":
# logging.info("Current Directory 3")
@ -196,7 +196,7 @@ async def _document_loader(observation: str, loader_settings: dict):
# pages = chunk_data(chunk_strategy=loader_strategy, source_data=str(documents), chunk_size=chunk_size,
# chunk_overlap=chunk_overlap)
# logging.info("Documents: %s", documents)
# from cognitive_architecture.shared.language_processing import translate_text, detect_language
# from cognee.shared.language_processing import translate_text, detect_language
#
# if detect_language(pages) != "en":
# logging.info("Current Directory 3")

View file

@ -3,7 +3,7 @@ import logging
from langchain.text_splitter import RecursiveCharacterTextSplitter
from marshmallow import Schema, fields
from cognitive_architecture.database.vectordb.loaders.loaders import _document_loader
from cognee.database.vectordb.loaders.loaders import _document_loader
# Add the parent directory to sys.path
@ -186,7 +186,7 @@ class WeaviateVectorDB(VectorDB):
)
else:
chunk_count = 0
from cognitive_architecture.database.vectordb.chunkers.chunkers import (
from cognee.database.vectordb.chunkers.chunkers import (
chunk_data,
)

View file

@ -1,4 +1,4 @@
from cognitive_architecture.config import Config
from cognee.config import Config
from .databases.relational import SqliteEngine, DatabaseEngine
config = Config()

View file

@ -1,8 +1,8 @@
import logging
from . import Dataset, Data
from cognitive_architecture.infrastructure import infrastructure_config
from cognitive_architecture.infrastructure.databases.relational import DatabaseEngine
from cognitive_architecture.infrastructure.files import remove_file_from_storage
from cognee.infrastructure import infrastructure_config
from cognee.infrastructure.databases.relational import DatabaseEngine
from cognee.infrastructure.files import remove_file_from_storage
logger = logging.getLogger(__name__)

View file

@ -2,7 +2,7 @@ from typing import List
from datetime import datetime, timezone
from sqlalchemy.orm import relationship, MappedColumn, Mapped
from sqlalchemy import Column, String, DateTime, UUID, Text, JSON
from cognitive_architecture.infrastructure.databases.relational import ModelBase
from cognee.infrastructure.databases.relational import ModelBase
from .DatasetData import DatasetData
class Data(ModelBase):

View file

@ -2,7 +2,7 @@ from typing import List
from datetime import datetime, timezone
from sqlalchemy.orm import relationship, Mapped
from sqlalchemy import Column, Text, DateTime, UUID
from cognitive_architecture.infrastructure.databases.relational import ModelBase
from cognee.infrastructure.databases.relational import ModelBase
from .DatasetData import DatasetData
class Dataset(ModelBase):

View file

@ -1,7 +1,7 @@
from uuid import uuid4
from datetime import datetime, timezone
from sqlalchemy import Column, DateTime, UUID, ForeignKey
from cognitive_architecture.infrastructure.databases.relational import ModelBase
from cognee.infrastructure.databases.relational import ModelBase
class DatasetData(ModelBase):
__tablename__ = "dataset_data"

View file

@ -1,7 +1,7 @@
"""Factory function to get the appropriate graph client based on the graph type."""
from cognitive_architecture.config import Config
from cognitive_architecture.root_dir import get_absolute_path
from cognitive_architecture.shared.data_models import GraphDBType
from cognee.config import Config
from cognee.root_dir import get_absolute_path
from cognee.shared.data_models import GraphDBType
from .graph_db_interface import GraphDBInterface
from .networkx.adapter import NetworXAdapter
# from .neo4j.adapter import Neo4jAdapter

View file

@ -8,7 +8,7 @@ from typing import Optional, Dict, Any
import aiofiles.os
import aiofiles
import networkx as nx
from cognitive_architecture.infrastructure.databases.graph.graph_db_interface import GraphDBInterface
from cognee.infrastructure.databases.graph.graph_db_interface import GraphDBInterface
import logging
class NetworXAdapter(GraphDBInterface):

View file

@ -1,5 +1,5 @@
import duckdb
from cognitive_architecture.root_dir import get_absolute_path
from cognee.root_dir import get_absolute_path
class DuckDBAdapter():
def __init__(self):

View file

@ -2,9 +2,9 @@ import uuid
from pathlib import Path
from sqlalchemy import select
from sqlalchemy.ext.asyncio import create_async_engine, async_sessionmaker, AsyncSession
from cognitive_architecture.config import Config
from cognee.config import Config
# from ..relational_db_interface import RelationalDBInterface
from cognitive_architecture.database.relationaldb.models.memory import MemoryModel
from cognee.database.relationaldb.models.memory import MemoryModel
config = Config()
config.load()

View file

@ -2,7 +2,7 @@
# from sqlalchemy.orm import relationship
# # from sqlalchemy.orm import DeclarativeBase
# from sqlalchemy import Column, String, DateTime, ForeignKey
# from cognitive_architecture.database.relationaldb.database import Base
# from cognee.database.relationaldb.database import Base
# class MemoryModel(Base):

View file

@ -4,7 +4,7 @@ from typing import Callable
from sqlalchemy.inspection import inspect
from sqlalchemy.ext.asyncio import create_async_engine, async_sessionmaker, AsyncEngine, AsyncSession, async_scoped_session
from sqlalchemy.future import select
from cognitive_architecture.infrastructure.files.storage.LocalStorage import LocalStorage
from cognee.infrastructure.files.storage.LocalStorage import LocalStorage
from ..DatabaseEngine import DatabaseEngine
from ..ModelBase import ModelBase
from ..utils import with_rollback

View file

@ -1,4 +1,4 @@
from cognitive_architecture.config import Config
from cognee.config import Config
from .qdrant import QDrantAdapter
config = Config()

View file

@ -3,7 +3,7 @@ from langchain.embeddings.openai import OpenAIEmbeddings
from langchain.retrievers import WeaviateHybridSearchRetriever, ParentDocumentRetriever
from databases.vector.vector_db_interface import VectorDBInterface
# from langchain.text_splitter import RecursiveCharacterTextSplitter
from cognitive_architecture.database.vectordb.loaders.loaders import _document_loader
from cognee.database.vectordb.loaders.loaders import _document_loader
class WeaviateVectorDB(VectorDBInterface):
def __init__(self, *args, **kwargs):
@ -136,7 +136,7 @@ class WeaviateVectorDB(VectorDBInterface):
)
else:
chunk_count = 0
from cognitive_architecture.database.vectordb.chunkers.chunkers import (
from cognee.database.vectordb.chunkers.chunkers import (
chunk_data,
)

View file

@ -1,5 +1,5 @@
from typing import BinaryIO
from cognitive_architecture.root_dir import get_absolute_path
from cognee.root_dir import get_absolute_path
from .storage.StorageManager import StorageManager
from .storage.LocalStorage import LocalStorage

View file

@ -1,4 +1,4 @@
from cognitive_architecture.root_dir import get_absolute_path
from cognee.root_dir import get_absolute_path
from .storage.StorageManager import StorageManager
from .storage.LocalStorage import LocalStorage

View file

@ -1,5 +1,5 @@
"""Get the LLM client."""
from cognitive_architecture.config import Config
from cognee.config import Config
from .openai.adapter import OpenAIAdapter
config = Config()

View file

@ -6,7 +6,7 @@ import instructor
from openai import AsyncOpenAI
from pydantic import BaseModel
from tenacity import retry, stop_after_attempt
from cognitive_architecture.utils import read_query_prompt
from cognee.utils import read_query_prompt
from ..llm_interface import LLMInterface
class OpenAIAdapter(LLMInterface):

View file

@ -3,8 +3,8 @@ import os
import instructor
from openai import OpenAI
import logging
from cognitive_architecture.shared.data_models import KnowledgeGraph, MemorySummary
from cognitive_architecture.config import Config
from cognee.shared.data_models import KnowledgeGraph, MemorySummary
from cognee.config import Config
@ -36,7 +36,7 @@ def generate_graph(input) -> KnowledgeGraph:
model = "gpt-4-1106-preview"
user_prompt = f"Use the given format to extract information from the following input: {input}."
system_prompt = read_query_prompt(
"cognitive_architecture/llm/prompts/generate_graph_prompt.txt"
"cognee/llm/prompts/generate_graph_prompt.txt"
)
out = aclient.chat.completions.create(
@ -87,7 +87,7 @@ async def generate_summary(input) -> MemorySummary:
def user_query_to_edges_and_nodes(input: str) -> KnowledgeGraph:
"""Generate a knowledge graph from a user query."""
system_prompt = read_query_prompt(
"cognitive_architecture/llm/prompts/generate_graph_prompt.txt"
"cognee/llm/prompts/generate_graph_prompt.txt"
)
return aclient.chat.completions.create(
model=config.model,

View file

@ -1,7 +1,7 @@
from datetime import datetime
from sqlalchemy.orm import Mapped, MappedColumn
from sqlalchemy import Column, String, DateTime, ForeignKey, Enum, UUID, JSON
from cognitive_architecture.infrastructure.databases.relational import ModelBase
from cognee.infrastructure.databases.relational import ModelBase
class OperationType(Enum):
MERGE_DATA = "MERGE_DATA"

View file

@ -1,5 +1,5 @@
""" Here we update semantic graph with content that classifier produced"""
from cognitive_architecture.infrastructure.databases.graph.get_graph_client import get_graph_client, GraphDBType
from cognee.infrastructure.databases.graph.get_graph_client import get_graph_client, GraphDBType
async def add_classification_nodes(document_id, classification_data):

View file

@ -1,5 +1,5 @@
from cognitive_architecture.infrastructure.databases.graph.get_graph_client import get_graph_client
from cognitive_architecture.shared.data_models import GraphDBType, Document, DocumentType, Category, Relationship
from cognee.infrastructure.databases.graph.get_graph_client import get_graph_client
from cognee.shared.data_models import GraphDBType, Document, DocumentType, Category, Relationship
from .create import add_node_and_edge
def create_category(category_name: str):

View file

@ -1,6 +1,6 @@
from networkx import Graph
from cognitive_architecture.infrastructure.databases.graph.get_graph_client import get_graph_client
from cognitive_architecture.shared.data_models import GraphDBType
from cognee.infrastructure.databases.graph.get_graph_client import get_graph_client
from cognee.shared.data_models import GraphDBType
async def extract_node_descriptions(data):

View file

@ -2,7 +2,7 @@
import uuid
import json
from datetime import datetime
from cognitive_architecture.infrastructure.databases.graph.get_graph_client import get_graph_client, GraphDBType
from cognee.infrastructure.databases.graph.get_graph_client import get_graph_client, GraphDBType
async def add_propositions(

View file

@ -1,8 +1,8 @@
""" This module is responsible for creating a semantic graph """
from typing import Optional, Any
from pydantic import BaseModel
from cognitive_architecture.infrastructure.databases.graph.get_graph_client import get_graph_client
from cognitive_architecture.shared.data_models import GraphDBType
from cognee.infrastructure.databases.graph.get_graph_client import get_graph_client
from cognee.shared.data_models import GraphDBType
async def generate_node_id(instance: BaseModel) -> str:

View file

@ -1,6 +1,6 @@
from datetime import datetime
from cognitive_architecture.shared.data_models import DefaultGraphModel, Relationship, UserProperties, UserLocation
from cognitive_architecture.modules.cognify.graph.create import create_semantic_graph
from cognee.shared.data_models import DefaultGraphModel, Relationship, UserProperties, UserLocation
from cognee.modules.cognify.graph.create import create_semantic_graph
async def initialize_graph(root_id: str):
graph = DefaultGraphModel(

View file

@ -1,8 +1,8 @@
""" This module contains the code to classify content into categories using the LLM API. """
from typing import Type, List
from pydantic import BaseModel
from cognitive_architecture.infrastructure.llm.get_llm_client import get_llm_client
from cognitive_architecture.utils import read_query_prompt
from cognee.infrastructure.llm.get_llm_client import get_llm_client
from cognee.utils import read_query_prompt
async def classify_into_categories(text_input: str, system_prompt_path: str, response_model: Type[BaseModel]):
llm_client = get_llm_client()

View file

@ -1,8 +1,8 @@
""" This module contains the code to classify content into categories using the LLM API. """
from typing import Type
from pydantic import BaseModel
from cognitive_architecture.infrastructure.llm.get_llm_client import get_llm_client
from cognitive_architecture.utils import async_render_template
from cognee.infrastructure.llm.get_llm_client import get_llm_client
from cognee.utils import async_render_template
async def content_to_cog_layers(filename: str, context, response_model: Type[BaseModel]):
llm_client = get_llm_client()

View file

@ -2,9 +2,9 @@
import json
from typing import Type
from pydantic import BaseModel
from cognitive_architecture.infrastructure.llm.get_llm_client import get_llm_client
from cognitive_architecture.shared.data_models import KnowledgeGraph
from cognitive_architecture.utils import async_render_template
from cognee.infrastructure.llm.get_llm_client import get_llm_client
from cognee.shared.data_models import KnowledgeGraph
from cognee.utils import async_render_template
async def generate_graph(text_input: str, filename: str, context, response_model: Type[BaseModel]):
llm_client = get_llm_client()

View file

@ -1,8 +1,8 @@
import asyncio
from qdrant_client import models
from cognitive_architecture.infrastructure.llm.get_llm_client import get_llm_client
from cognitive_architecture.infrastructure.databases.vector import get_vector_database
from cognee.infrastructure.llm.get_llm_client import get_llm_client
from cognee.infrastructure.databases.vector import get_vector_database
async def get_embeddings(texts:list):
""" Get embeddings for a list of texts"""

View file

@ -1,5 +1,5 @@
from typing import BinaryIO
from cognitive_architecture.infrastructure.files import get_file_metadata, FileMetadata
from cognee.infrastructure.files import get_file_metadata, FileMetadata
from .IngestionData import IngestionData
def create_binary_data(data: BinaryIO):

View file

@ -1,7 +1,7 @@
import asyncio
from uuid import UUID, uuid4
from cognitive_architecture.infrastructure.files import add_file_to_storage
from cognitive_architecture.infrastructure.data import add_data_to_dataset, Data, Dataset
from cognee.infrastructure.files import add_file_to_storage
from cognee.infrastructure.data import add_data_to_dataset, Data, Dataset
from .data_types import IngestionData
async def save(dataset_id: UUID, dataset_name: str, data_id: UUID, data: IngestionData):

View file

@ -1,5 +1,5 @@
from cognitive_architecture.infrastructure.databases.vector.qdrant.adapter import CollectionConfig
from cognitive_architecture.infrastructure.databases.vector.get_vector_database import get_vector_database
from cognee.infrastructure.databases.vector.qdrant.adapter import CollectionConfig
from cognee.infrastructure.databases.vector.get_vector_database import get_vector_database
async def create_vector_memory(memory_name: str, collection_config: CollectionConfig):
vector_db = get_vector_database()

View file

@ -1,5 +1,5 @@
""" Fetches the context of a given node in the graph"""
from cognitive_architecture.infrastructure.databases.graph.get_graph_client import get_graph_client
from cognee.infrastructure.databases.graph.get_graph_client import get_graph_client
async def search_neighbour(CONNECTED_GRAPH, id):
relevant_context = []
for n,attr in CONNECTED_GRAPH.nodes(data=True):
@ -18,7 +18,7 @@ async def search_neighbour(CONNECTED_GRAPH, id):
if __name__ == '__main__':
import asyncio
async def main():
from cognitive_architecture.shared.data_models import GraphDBType
from cognee.shared.data_models import GraphDBType
graph_client = get_graph_client(GraphDBType.NETWORKX)
graph = await graph_client.graph

View file

@ -1,6 +1,6 @@
from cognitive_architecture.infrastructure.llm.get_llm_client import get_llm_client
from cognitive_architecture.modules.cognify.graph.add_node_connections import extract_node_descriptions
from cognee.infrastructure.llm.get_llm_client import get_llm_client
from cognee.modules.cognify.graph.add_node_connections import extract_node_descriptions
async def search_similarity(query:str ,graph):
@ -15,7 +15,7 @@ async def search_similarity(query:str ,graph):
query = await client.async_get_embedding_with_backoff(query)
# print(query)
for id in unique_layer_uuids:
from cognitive_architecture.infrastructure.databases.vector.get_vector_database import get_vector_database
from cognee.infrastructure.databases.vector.get_vector_database import get_vector_database
vector_client = get_vector_database()
print(query)

View file

@ -1,8 +1,8 @@
import uuid
from typing import List
from qdrant_client.models import PointStruct
from cognitive_architecture.infrastructure.databases.vector.get_vector_database import get_vector_database
from cognitive_architecture.infrastructure.llm.openai.openai_tools import async_get_embedding_with_backoff
from cognee.infrastructure.databases.vector.get_vector_database import get_vector_database
from cognee.infrastructure.llm.openai.openai_tools import async_get_embedding_with_backoff
async def create_information_points(memory_name: str, payload: List[str]):
vector_db = get_vector_database()

View file

@ -1,4 +1,4 @@
from cognitive_architecture.infrastructure.databases.relational.get_database import get_database
from cognee.infrastructure.databases.relational.get_database import get_database
async def is_existing_memory(memory_name: str):
memory = await (get_database().get_memory_by_name(memory_name))

View file

@ -1,4 +1,4 @@
from cognitive_architecture.infrastructure.databases.relational.get_database import get_database
from cognee.infrastructure.databases.relational.get_database import get_database
def register_memory_for_user(user_id: str, memory_name: str):
return get_database().add_memory(user_id, memory_name)

View file

@ -5,11 +5,11 @@ logger = logging.getLogger(__name__)
async def main():
"""Runs as a part of startup docker scripts to create the database and tables."""
from cognitive_architecture.config import Config
from cognee.config import Config
config = Config()
config.load()
from cognitive_architecture.database.database_manager import DatabaseManager
from cognee.database.database_manager import DatabaseManager
db_manager = DatabaseManager()
database_name = config.db_name

View file

@ -12,11 +12,11 @@ from sqlalchemy import or_
from sqlalchemy.future import select
from sqlalchemy.orm import contains_eager
from sqlalchemy.ext.asyncio import AsyncSession
from cognitive_architecture.database.relationaldb.models.docs import DocsModel
from cognitive_architecture.database.relationaldb.models.memory import MemoryModel
from cognitive_architecture.database.relationaldb.models.operation import Operation
from cognee.database.relationaldb.models.docs import DocsModel
from cognee.database.relationaldb.models.memory import MemoryModel
from cognee.database.relationaldb.models.operation import Operation
from cognitive_architecture.config import Config
from cognee.config import Config
config = Config()
config.load()
@ -290,7 +290,7 @@ async def read_query_prompt(filename: str) -> str:
script_directory = Path(__file__).parent
# Set the base directory relative to the script's directory
base_directory = script_directory.parent / "cognitive_architecture/infrastructure/llm/prompts"
base_directory = script_directory.parent / "cognee/infrastructure/llm/prompts"
# Construct the full file path
file_path = base_directory / filename
@ -326,7 +326,7 @@ async def async_render_template(filename: str, context: dict) -> str:
script_directory = Path(__file__).parent
# Set the base directory relative to the script's directory
base_directory = script_directory.parent / "cognitive_architecture/infrastructure/llm/prompts"
base_directory = script_directory.parent / "cognee/infrastructure/llm/prompts"
# Construct the full file path

View file

@ -3,15 +3,15 @@ import logging
logging.basicConfig(level=logging.INFO)
from sqlalchemy.future import select
from cognitive_architecture.database.relationaldb.models.user import User
from cognitive_architecture.database.relationaldb.models.memory import MemoryModel
from cognee.database.relationaldb.models.user import User
from cognee.database.relationaldb.models.memory import MemoryModel
import ast
import tracemalloc
from cognitive_architecture.database.relationaldb.database_crud import add_entity
from cognee.database.relationaldb.database_crud import add_entity
tracemalloc.start()
import uuid
from cognitive_architecture.database.vectordb.basevectordb import BaseMemory
from cognitive_architecture.config import Config
from cognee.database.vectordb.basevectordb import BaseMemory
from cognee.config import Config
globalConfig = Config()

View file

@ -7,7 +7,7 @@ echo "Environment: $ENVIRONMENT"
if [ "$ENVIRONMENT" != "local" ]; then
echo "Running fetch_secret.py"
PYTHONPATH=. python cognitive_architecture/fetch_secret.py
PYTHONPATH=. python cognee/fetch_secret.py
if [ $? -ne 0 ]; then
echo "Error: fetch_secret.py failed"
@ -19,7 +19,7 @@ fi
echo "Creating database..."
PYTHONPATH=. python cognitive_architecture/setup_database.py
PYTHONPATH=. python cognee/setup_database.py
if [ $? -ne 0 ]; then
echo "Error: setup_database.py failed"
exit 1

46
main.py
View file

@ -2,53 +2,53 @@ from typing import Optional, List
from neo4j.exceptions import Neo4jError
from pydantic import BaseModel, Field
from cognitive_architecture.database.graphdb.graph import Neo4jGraphDB
from cognitive_architecture.database.relationaldb.models.memory import MemoryModel
from cognee.database.graphdb.graph import Neo4jGraphDB
from cognee.database.relationaldb.models.memory import MemoryModel
import os
from cognitive_architecture.database.relationaldb.database_crud import (
from cognee.database.relationaldb.database_crud import (
session_scope,
update_entity_graph_summary,
)
from cognitive_architecture.database.relationaldb.database import AsyncSessionLocal
from cognitive_architecture.utils import generate_letter_uuid
from cognee.database.relationaldb.database import AsyncSessionLocal
from cognee.utils import generate_letter_uuid
import instructor
from openai import OpenAI
from cognitive_architecture.vectorstore_manager import Memory
from cognitive_architecture.database.relationaldb.database_crud import fetch_job_id
from cognee.vectorstore_manager import Memory
from cognee.database.relationaldb.database_crud import fetch_job_id
import uuid
from cognitive_architecture.database.relationaldb.models.sessions import Session
from cognitive_architecture.database.relationaldb.models.operation import Operation
from cognitive_architecture.database.relationaldb.database_crud import (
from cognee.database.relationaldb.models.sessions import Session
from cognee.database.relationaldb.models.operation import Operation
from cognee.database.relationaldb.database_crud import (
session_scope,
add_entity,
update_entity,
fetch_job_id,
)
from cognitive_architecture.database.relationaldb.models.metadatas import MetaDatas
from cognitive_architecture.database.relationaldb.models.docs import DocsModel
from cognitive_architecture.database.relationaldb.models.memory import MemoryModel
from cognitive_architecture.database.relationaldb.models.user import User
from cognitive_architecture.classifiers.classify_summary import classify_summary
from cognitive_architecture.classifiers.classify_documents import classify_documents
from cognitive_architecture.classifiers.classify_user_query import classify_user_query
from cognitive_architecture.classifiers.classify_user_input import classify_user_input
from cognee.database.relationaldb.models.metadatas import MetaDatas
from cognee.database.relationaldb.models.docs import DocsModel
from cognee.database.relationaldb.models.memory import MemoryModel
from cognee.database.relationaldb.models.user import User
from cognee.classifiers.classify_summary import classify_summary
from cognee.classifiers.classify_documents import classify_documents
from cognee.classifiers.classify_user_query import classify_user_query
from cognee.classifiers.classify_user_input import classify_user_input
aclient = instructor.patch(OpenAI())
DEFAULT_PRESET = "promethai_chat"
preset_options = [DEFAULT_PRESET]
PROMETHAI_DIR = os.path.join(os.path.expanduser("~"), ".")
OPENAI_API_KEY = os.getenv("OPENAI_API_KEY", "")
from cognitive_architecture.config import Config
from cognee.config import Config
config = Config()
config.load()
from cognitive_architecture.utils import get_document_names
from cognee.utils import get_document_names
from sqlalchemy.orm import selectinload, joinedload, contains_eager
import logging
from sqlalchemy.ext.asyncio import AsyncSession
from sqlalchemy.future import select
from cognitive_architecture.utils import (
from cognee.utils import (
get_document_names,
generate_letter_uuid,
get_memory_name_by_doc_id,
@ -56,7 +56,7 @@ from cognitive_architecture.utils import (
get_vectordb_namespace,
get_vectordb_document_name,
)
from cognitive_architecture.shared.language_processing import (
from cognee.shared.language_processing import (
translate_text,
detect_language,
)
@ -152,7 +152,7 @@ async def load_documents_to_vectorstore(
# except:
# document_names = document_source
for doc in document_names:
from cognitive_architecture.shared.language_processing import (
from cognee.shared.language_processing import (
translate_text,
detect_language,
)

6
poetry.lock generated
View file

@ -4096,7 +4096,6 @@ description = "Nvidia JIT LTO Library"
optional = false
python-versions = ">=3"
files = [
{file = "nvidia_nvjitlink_cu12-12.4.99-py3-none-manylinux2014_aarch64.whl", hash = "sha256:75d6498c96d9adb9435f2bbdbddb479805ddfb97b5c1b32395c694185c20ca57"},
{file = "nvidia_nvjitlink_cu12-12.4.99-py3-none-manylinux2014_x86_64.whl", hash = "sha256:c6428836d20fe7e327191c175791d38570e10762edc588fb46749217cd444c74"},
{file = "nvidia_nvjitlink_cu12-12.4.99-py3-none-win_amd64.whl", hash = "sha256:991905ffa2144cb603d8ca7962d75c35334ae82bf92820b6ba78157277da1ad2"},
]
@ -5701,6 +5700,7 @@ files = [
{file = "PyYAML-6.0.1-cp311-cp311-win_amd64.whl", hash = "sha256:bf07ee2fef7014951eeb99f56f39c9bb4af143d8aa3c21b1677805985307da34"},
{file = "PyYAML-6.0.1-cp312-cp312-macosx_10_9_x86_64.whl", hash = "sha256:855fb52b0dc35af121542a76b9a84f8d1cd886ea97c84703eaa6d88e37a2ad28"},
{file = "PyYAML-6.0.1-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:40df9b996c2b73138957fe23a16a4f0ba614f4c0efce1e9406a184b6d07fa3a9"},
{file = "PyYAML-6.0.1-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a08c6f0fe150303c1c6b71ebcd7213c2858041a7e01975da3a99aed1e7a378ef"},
{file = "PyYAML-6.0.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6c22bec3fbe2524cde73d7ada88f6566758a8f7227bfbf93a408a9d86bcc12a0"},
{file = "PyYAML-6.0.1-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:8d4e9c88387b0f5c7d5f281e55304de64cf7f9c0021a3525bd3b1c542da3b0e4"},
{file = "PyYAML-6.0.1-cp312-cp312-win32.whl", hash = "sha256:d483d2cdf104e7c9fa60c544d92981f12ad66a457afae824d146093b8c294c54"},
@ -8057,5 +8057,5 @@ weaviate = []
[metadata]
lock-version = "2.0"
python-versions = "3.10.13"
content-hash = "d9010f8123850150bf7a9cf1c2955a69c69ce28467bbe8b6ff9d7bf0c7f072e1"
python-versions = "~3.10"
content-hash = "48d8cbd3319b1370abddeb2ff6c32cf6be01bdaff150d466129a1667b8cce94f"

View file

@ -17,7 +17,7 @@ classifiers = [
"Operating System :: Microsoft :: Windows",]
[tool.poetry.dependencies]
python = "3.10.13"
python = "~3.10"
langchain = "^0.0.338"
openai = "1.12.0"
python-dotenv = "1.0.1"

File diff suppressed because it is too large Load diff