chore: rename package in files
This commit is contained in:
parent
40e760f517
commit
d5391f903c
66 changed files with 6022 additions and 6022 deletions
|
|
@ -7,7 +7,7 @@ POSTGRES_PASSWORD = bla
|
|||
POSTGRES_DB = bubu
|
||||
POSTGRES_HOST = localhost
|
||||
POSTGRES_HOST_DOCKER = postgres
|
||||
COG_ARCH_DIR = cognitive_architecture
|
||||
COG_ARCH_DIR = cognee
|
||||
GRAPH_DB_URL =
|
||||
GRAPH_DB_PW =
|
||||
GRAPH_DB_USER =
|
||||
|
|
|
|||
2
.gitignore
vendored
2
.gitignore
vendored
|
|
@ -164,4 +164,4 @@ cython_debug/
|
|||
|
||||
.vscode/
|
||||
database/data/
|
||||
cognitive_architecture/data/
|
||||
cognee/data/
|
||||
|
|
|
|||
9364
Demo_graph.ipynb
9364
Demo_graph.ipynb
File diff suppressed because one or more lines are too long
|
|
@ -43,7 +43,7 @@ WORKDIR /app
|
|||
# Set the PYTHONPATH environment variable to include the /app directory
|
||||
ENV PYTHONPATH=/app
|
||||
|
||||
COPY cognitive_architecture/ /app/cognitive_architecture
|
||||
COPY cognee/ /app/cognee
|
||||
COPY main.py /app
|
||||
COPY api.py /app
|
||||
|
||||
|
|
|
|||
20
api.py
20
api.py
|
|
@ -13,7 +13,7 @@ logging.basicConfig(
|
|||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
from cognitive_architecture.config import Config
|
||||
from cognee.config import Config
|
||||
|
||||
config = Config()
|
||||
config.load()
|
||||
|
|
@ -23,9 +23,9 @@ from fastapi import FastAPI, BackgroundTasks, HTTPException
|
|||
from fastapi.responses import JSONResponse
|
||||
from pydantic import BaseModel
|
||||
|
||||
from cognitive_architecture.database.relationaldb.database import AsyncSessionLocal
|
||||
from cognitive_architecture.database.relationaldb.database_crud import session_scope
|
||||
from cognitive_architecture.vectorstore_manager import Memory
|
||||
from cognee.database.relationaldb.database import AsyncSessionLocal
|
||||
from cognee.database.relationaldb.database_crud import session_scope
|
||||
from cognee.vectorstore_manager import Memory
|
||||
from main import add_documents_to_graph_db, user_context_enrichment
|
||||
|
||||
OPENAI_API_KEY = os.getenv("OPENAI_API_KEY", "")
|
||||
|
|
@ -57,7 +57,7 @@ def health_check():
|
|||
class Payload(BaseModel):
|
||||
payload: Dict[str, Any]
|
||||
|
||||
from cognitive_architecture.api.v1.memory.create_memory import MemoryType
|
||||
from cognee.api.v1.memory.create_memory import MemoryType
|
||||
|
||||
class CreateMemoryPayload(BaseModel):
|
||||
user_id: str
|
||||
|
|
@ -66,7 +66,7 @@ class CreateMemoryPayload(BaseModel):
|
|||
|
||||
@app.post("/create-memory", response_model=dict)
|
||||
async def create_memory(payload: CreateMemoryPayload):
|
||||
from cognitive_architecture.api.v1.memory.create_memory import create_memory as create_memory_v1, MemoryException
|
||||
from cognee.api.v1.memory.create_memory import create_memory as create_memory_v1, MemoryException
|
||||
|
||||
try:
|
||||
await create_memory_v1(
|
||||
|
|
@ -93,7 +93,7 @@ class RememberPayload(BaseModel):
|
|||
|
||||
@app.post("/remember", response_model=dict)
|
||||
async def remember(payload: RememberPayload):
|
||||
from cognitive_architecture.api.v1.memory.remember import remember as remember_v1
|
||||
from cognee.api.v1.memory.remember import remember as remember_v1
|
||||
|
||||
await remember_v1(
|
||||
payload.user_id,
|
||||
|
|
@ -264,7 +264,7 @@ async def user_query_classfier(payload: Payload):
|
|||
|
||||
# Execute the query - replace this with the actual execution method
|
||||
async with session_scope(session=AsyncSessionLocal()) as session:
|
||||
from cognitive_architecture.classifiers.classify_user_input import (
|
||||
from cognee.classifiers.classify_user_input import (
|
||||
classify_user_query,
|
||||
)
|
||||
|
||||
|
|
@ -292,7 +292,7 @@ async def drop_db(payload: Payload):
|
|||
else:
|
||||
pass
|
||||
|
||||
from cognitive_architecture.database.create_database import (
|
||||
from cognee.database.create_database import (
|
||||
drop_database,
|
||||
create_admin_engine,
|
||||
)
|
||||
|
|
@ -310,7 +310,7 @@ async def drop_db(payload: Payload):
|
|||
else:
|
||||
pass
|
||||
|
||||
from cognitive_architecture.database.create_database import (
|
||||
from cognee.database.create_database import (
|
||||
create_database,
|
||||
create_admin_engine,
|
||||
)
|
||||
|
|
|
|||
File diff suppressed because one or more lines are too long
184
cognee.ipynb
184
cognee.ipynb
|
|
@ -1,94 +1,94 @@
|
|||
{
|
||||
"cells": [
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "c4d5a399",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"from os import listdir, path\n",
|
||||
"from uuid import uuid5, UUID\n",
|
||||
"from cognitive_architecture import add\n",
|
||||
"\n",
|
||||
"data_path = path.abspath(\".data\")\n",
|
||||
"\n",
|
||||
"results = await add(data_path, \"izmene\")\n",
|
||||
"for result in results:\n",
|
||||
" print(result)\n"
|
||||
]
|
||||
"cells": [
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "c4d5a399",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"from os import listdir, path\n",
|
||||
"from uuid import uuid5, UUID\n",
|
||||
"from cognee import add\n",
|
||||
"\n",
|
||||
"data_path = path.abspath(\".data\")\n",
|
||||
"\n",
|
||||
"results = await add(data_path, \"izmene\")\n",
|
||||
"for result in results:\n",
|
||||
" print(result)\n"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "47edce91",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"import duckdb\n",
|
||||
"from cognee.root_dir import get_absolute_path\n",
|
||||
"\n",
|
||||
"dataset_name = \"pdf_files\"\n",
|
||||
"\n",
|
||||
"db_path = get_absolute_path(\"./data/cognee\")\n",
|
||||
"db_location = db_path + \"/cognee.duckdb\"\n",
|
||||
"print(db_location)\n",
|
||||
"\n",
|
||||
"db = duckdb.connect(db_location)\n",
|
||||
"\n",
|
||||
"tables = db.sql(\"SELECT DISTINCT schema_name FROM duckdb_tables();\").df()\n",
|
||||
"print(list(filter(lambda table_name: table_name.endswith('staging') is False, tables.to_dict()[\"schema_name\"].values())))\n",
|
||||
"\n",
|
||||
"# izmene = db.sql(f\"SELECT * FROM izmene.file_metadata;\")\n",
|
||||
"\n",
|
||||
"# print(izmene)\n",
|
||||
"\n",
|
||||
"# pravilnik = db.sql(f\"SELECT * FROM pravilnik.file_metadata;\")\n",
|
||||
"\n",
|
||||
"# print(pravilnik)\n",
|
||||
"\n",
|
||||
"\n"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "607bf624",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"from os import path, listdir\n",
|
||||
"from cognee import cognify, list_datasets\n",
|
||||
"from cognee.utils import render_graph\n",
|
||||
"\n",
|
||||
"print(list_datasets())\n",
|
||||
"\n",
|
||||
"graph = await cognify(\"izmene\")\n",
|
||||
"\n",
|
||||
"graph_url = await render_graph(graph, graph_type = \"networkx\")\n",
|
||||
"print(graph_url)\n"
|
||||
]
|
||||
}
|
||||
],
|
||||
"metadata": {
|
||||
"kernelspec": {
|
||||
"display_name": "Python 3 (ipykernel)",
|
||||
"language": "python",
|
||||
"name": "python3"
|
||||
},
|
||||
"language_info": {
|
||||
"codemirror_mode": {
|
||||
"name": "ipython",
|
||||
"version": 3
|
||||
},
|
||||
"file_extension": ".py",
|
||||
"mimetype": "text/x-python",
|
||||
"name": "python",
|
||||
"nbconvert_exporter": "python",
|
||||
"pygments_lexer": "ipython3",
|
||||
"version": "3.10.13"
|
||||
}
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "47edce91",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"import duckdb\n",
|
||||
"from cognitive_architecture.root_dir import get_absolute_path\n",
|
||||
"\n",
|
||||
"dataset_name = \"pdf_files\"\n",
|
||||
"\n",
|
||||
"db_path = get_absolute_path(\"./data/cognee\")\n",
|
||||
"db_location = db_path + \"/cognee.duckdb\"\n",
|
||||
"print(db_location)\n",
|
||||
"\n",
|
||||
"db = duckdb.connect(db_location)\n",
|
||||
"\n",
|
||||
"tables = db.sql(\"SELECT DISTINCT schema_name FROM duckdb_tables();\").df()\n",
|
||||
"print(list(filter(lambda table_name: table_name.endswith('staging') is False, tables.to_dict()[\"schema_name\"].values())))\n",
|
||||
"\n",
|
||||
"# izmene = db.sql(f\"SELECT * FROM izmene.file_metadata;\")\n",
|
||||
"\n",
|
||||
"# print(izmene)\n",
|
||||
"\n",
|
||||
"# pravilnik = db.sql(f\"SELECT * FROM pravilnik.file_metadata;\")\n",
|
||||
"\n",
|
||||
"# print(pravilnik)\n",
|
||||
"\n",
|
||||
"\n"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "607bf624",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"from os import path, listdir\n",
|
||||
"from cognitive_architecture import cognify, list_datasets\n",
|
||||
"from cognitive_architecture.utils import render_graph\n",
|
||||
"\n",
|
||||
"print(list_datasets())\n",
|
||||
"\n",
|
||||
"graph = await cognify(\"izmene\")\n",
|
||||
"\n",
|
||||
"graph_url = await render_graph(graph, graph_type = \"networkx\")\n",
|
||||
"print(graph_url)\n"
|
||||
]
|
||||
}
|
||||
],
|
||||
"metadata": {
|
||||
"kernelspec": {
|
||||
"display_name": "Python 3 (ipykernel)",
|
||||
"language": "python",
|
||||
"name": "python3"
|
||||
},
|
||||
"language_info": {
|
||||
"codemirror_mode": {
|
||||
"name": "ipython",
|
||||
"version": 3
|
||||
},
|
||||
"file_extension": ".py",
|
||||
"mimetype": "text/x-python",
|
||||
"name": "python",
|
||||
"nbconvert_exporter": "python",
|
||||
"pygments_lexer": "ipython3",
|
||||
"version": "3.10.13"
|
||||
}
|
||||
},
|
||||
"nbformat": 4,
|
||||
"nbformat_minor": 5
|
||||
}
|
||||
"nbformat": 4,
|
||||
"nbformat_minor": 5
|
||||
}
|
||||
|
|
@ -4,10 +4,10 @@ import asyncio
|
|||
import dlt
|
||||
import duckdb
|
||||
from unstructured.cleaners.core import clean
|
||||
from cognitive_architecture.root_dir import get_absolute_path
|
||||
import cognitive_architecture.modules.ingestion as ingestion
|
||||
from cognitive_architecture.infrastructure.files import get_file_metadata
|
||||
from cognitive_architecture.infrastructure.files.storage import LocalStorage
|
||||
from cognee.root_dir import get_absolute_path
|
||||
import cognee.modules.ingestion as ingestion
|
||||
from cognee.infrastructure.files import get_file_metadata
|
||||
from cognee.infrastructure.files.storage import LocalStorage
|
||||
|
||||
async def add(file_paths: Union[str, List[str]], dataset_name: str = None):
|
||||
if isinstance(file_paths, str):
|
||||
|
|
|
|||
|
|
@ -1,8 +1,8 @@
|
|||
import asyncio
|
||||
from uuid import UUID, uuid4
|
||||
from typing import Union, BinaryIO, List
|
||||
import cognitive_architecture.modules.ingestion as ingestion
|
||||
from cognitive_architecture.infrastructure import infrastructure_config
|
||||
import cognee.modules.ingestion as ingestion
|
||||
from cognee.infrastructure import infrastructure_config
|
||||
|
||||
class DatasetException(Exception):
|
||||
message: str
|
||||
|
|
|
|||
|
|
@ -1,6 +1,6 @@
|
|||
from typing import List
|
||||
from enum import Enum
|
||||
from cognitive_architecture.modules.users.memory import create_information_points, is_existing_memory
|
||||
from cognee.modules.users.memory import create_information_points, is_existing_memory
|
||||
|
||||
class MemoryType(Enum):
|
||||
GRAPH = "GRAPH"
|
||||
|
|
|
|||
|
|
@ -6,27 +6,27 @@ import instructor
|
|||
from openai import OpenAI
|
||||
from unstructured.cleaners.core import clean
|
||||
from unstructured.partition.pdf import partition_pdf
|
||||
from cognitive_architecture.infrastructure.databases.vector.qdrant.QDrantAdapter import CollectionConfig
|
||||
from cognitive_architecture.infrastructure.llm.get_llm_client import get_llm_client
|
||||
from cognitive_architecture.modules.cognify.graph.add_classification_nodes import add_classification_nodes
|
||||
from cognitive_architecture.modules.cognify.graph.add_node_connections import add_node_connection, graph_ready_output, \
|
||||
from cognee.infrastructure.databases.vector.qdrant.QDrantAdapter import CollectionConfig
|
||||
from cognee.infrastructure.llm.get_llm_client import get_llm_client
|
||||
from cognee.modules.cognify.graph.add_classification_nodes import add_classification_nodes
|
||||
from cognee.modules.cognify.graph.add_node_connections import add_node_connection, graph_ready_output, \
|
||||
connect_nodes_in_graph, extract_node_descriptions
|
||||
from cognitive_architecture.modules.cognify.graph.add_propositions import append_to_graph
|
||||
from cognitive_architecture.modules.cognify.llm.add_node_connection_embeddings import process_items
|
||||
from cognitive_architecture.modules.cognify.vector.batch_search import adapted_qdrant_batch_search
|
||||
from cognitive_architecture.modules.cognify.vector.add_propositions import add_propositions
|
||||
from cognee.modules.cognify.graph.add_propositions import append_to_graph
|
||||
from cognee.modules.cognify.llm.add_node_connection_embeddings import process_items
|
||||
from cognee.modules.cognify.vector.batch_search import adapted_qdrant_batch_search
|
||||
from cognee.modules.cognify.vector.add_propositions import add_propositions
|
||||
|
||||
from cognitive_architecture.config import Config
|
||||
from cognitive_architecture.modules.cognify.llm.classify_content import classify_into_categories
|
||||
from cognitive_architecture.modules.cognify.llm.content_to_cog_layers import content_to_cog_layers
|
||||
from cognitive_architecture.modules.cognify.llm.generate_graph import generate_graph
|
||||
from cognitive_architecture.shared.data_models import DefaultContentPrediction, KnowledgeGraph, DefaultCognitiveLayer
|
||||
from cognitive_architecture.infrastructure.databases.graph.get_graph_client import get_graph_client
|
||||
from cognitive_architecture.shared.data_models import GraphDBType
|
||||
from cognitive_architecture.infrastructure.databases.vector.get_vector_database import get_vector_database
|
||||
from cognitive_architecture.infrastructure.databases.relational import DuckDBAdapter
|
||||
from cognitive_architecture.modules.cognify.graph.add_document_node import add_document_node
|
||||
from cognitive_architecture.modules.cognify.graph.initialize_graph import initialize_graph
|
||||
from cognee.config import Config
|
||||
from cognee.modules.cognify.llm.classify_content import classify_into_categories
|
||||
from cognee.modules.cognify.llm.content_to_cog_layers import content_to_cog_layers
|
||||
from cognee.modules.cognify.llm.generate_graph import generate_graph
|
||||
from cognee.shared.data_models import DefaultContentPrediction, KnowledgeGraph, DefaultCognitiveLayer
|
||||
from cognee.infrastructure.databases.graph.get_graph_client import get_graph_client
|
||||
from cognee.shared.data_models import GraphDBType
|
||||
from cognee.infrastructure.databases.vector.get_vector_database import get_vector_database
|
||||
from cognee.infrastructure.databases.relational import DuckDBAdapter
|
||||
from cognee.modules.cognify.graph.add_document_node import add_document_node
|
||||
from cognee.modules.cognify.graph.initialize_graph import initialize_graph
|
||||
|
||||
config = Config()
|
||||
config.load()
|
||||
|
|
|
|||
|
|
@ -1,5 +1,5 @@
|
|||
|
||||
from cognitive_architecture.infrastructure.databases.relational import DuckDBAdapter
|
||||
from cognee.infrastructure.databases.relational import DuckDBAdapter
|
||||
|
||||
def list_datasets():
|
||||
db = DuckDBAdapter()
|
||||
|
|
|
|||
|
|
@ -2,12 +2,12 @@
|
|||
from enum import Enum, auto
|
||||
from typing import Dict, Any, Callable, List
|
||||
|
||||
from cognitive_architecture.infrastructure.databases.graph.get_graph_client import get_graph_client
|
||||
from cognitive_architecture.modules.search.graph.search_adjacent import search_adjacent
|
||||
from cognitive_architecture.modules.search.vector.search_similarity import search_similarity
|
||||
from cognitive_architecture.modules.search.graph.search_categories import search_categories
|
||||
from cognitive_architecture.modules.search.graph.search_neighbour import search_neighbour
|
||||
from cognitive_architecture.shared.data_models import GraphDBType
|
||||
from cognee.infrastructure.databases.graph.get_graph_client import get_graph_client
|
||||
from cognee.modules.search.graph.search_adjacent import search_adjacent
|
||||
from cognee.modules.search.vector.search_similarity import search_similarity
|
||||
from cognee.modules.search.graph.search_categories import search_categories
|
||||
from cognee.modules.search.graph.search_neighbour import search_neighbour
|
||||
from cognee.shared.data_models import GraphDBType
|
||||
|
||||
|
||||
class SearchType(Enum):
|
||||
|
|
|
|||
|
|
@ -4,8 +4,8 @@ import os
|
|||
from contextlib import asynccontextmanager
|
||||
from sqlalchemy import text
|
||||
from sqlalchemy.ext.asyncio import create_async_engine
|
||||
from cognitive_architecture.config import Config
|
||||
from cognitive_architecture.database.relationaldb.database import Base, get_sqlalchemy_database_url
|
||||
from cognee.config import Config
|
||||
from cognee.database.relationaldb.database import Base, get_sqlalchemy_database_url
|
||||
|
||||
globalConfig = Config()
|
||||
|
||||
|
|
|
|||
|
|
@ -21,7 +21,7 @@ from ...utils import (
|
|||
create_edge_variable_mapping,
|
||||
create_node_variable_mapping,
|
||||
)
|
||||
from cognitive_architecture.infrastructure.llm.openai.queries import generate_summary, generate_graph
|
||||
from cognee.infrastructure.llm.openai.queries import generate_summary, generate_graph
|
||||
import logging
|
||||
from neo4j import AsyncGraphDatabase
|
||||
from contextlib import asynccontextmanager
|
||||
|
|
|
|||
|
|
@ -1,6 +1,6 @@
|
|||
import pickle
|
||||
from pathlib import Path
|
||||
from cognitive_architecture.config import Config
|
||||
from cognee.config import Config
|
||||
import networkx as nx
|
||||
config = Config()
|
||||
config = config.load()
|
||||
|
|
|
|||
|
|
@ -3,7 +3,7 @@ from pathlib import Path
|
|||
# from contextlib import asynccontextmanager
|
||||
from sqlalchemy.ext.asyncio import create_async_engine, async_sessionmaker, AsyncSession
|
||||
from sqlalchemy.orm import declarative_base
|
||||
from cognitive_architecture.config import Config
|
||||
from cognee.config import Config
|
||||
|
||||
globalConfig = Config()
|
||||
|
||||
|
|
|
|||
|
|
@ -16,12 +16,12 @@ from langchain.retrievers import WeaviateHybridSearchRetriever
|
|||
from weaviate.gql.get import HybridFusion
|
||||
|
||||
|
||||
from cognitive_architecture.database.relationaldb.models.sessions import Session
|
||||
from cognitive_architecture.database.relationaldb.models.metadatas import MetaDatas
|
||||
from cognitive_architecture.database.relationaldb.models.operation import Operation
|
||||
from cognitive_architecture.database.relationaldb.models.docs import DocsModel
|
||||
from cognee.database.relationaldb.models.sessions import Session
|
||||
from cognee.database.relationaldb.models.metadatas import MetaDatas
|
||||
from cognee.database.relationaldb.models.operation import Operation
|
||||
from cognee.database.relationaldb.models.docs import DocsModel
|
||||
from sqlalchemy.orm import sessionmaker
|
||||
from cognitive_architecture.database.relationaldb.database import engine
|
||||
from cognee.database.relationaldb.database import engine
|
||||
|
||||
from typing import Optional
|
||||
import time
|
||||
|
|
@ -31,7 +31,7 @@ tracemalloc.start()
|
|||
|
||||
from datetime import datetime
|
||||
from langchain.embeddings.openai import OpenAIEmbeddings
|
||||
from cognitive_architecture.database.vectordb.vectordb import (
|
||||
from cognee.database.vectordb.vectordb import (
|
||||
PineconeVectorDB,
|
||||
WeaviateVectorDB,
|
||||
LanceDB,
|
||||
|
|
@ -41,7 +41,7 @@ import uuid
|
|||
import weaviate
|
||||
from marshmallow import Schema, fields
|
||||
import json
|
||||
from cognitive_architecture.database.vectordb.vector_db_type import VectorDBType
|
||||
from cognee.database.vectordb.vector_db_type import VectorDBType
|
||||
|
||||
|
||||
OPENAI_API_KEY = os.getenv("OPENAI_API_KEY", "")
|
||||
|
|
|
|||
|
|
@ -2,7 +2,7 @@
|
|||
|
||||
import re
|
||||
import logging
|
||||
from cognitive_architecture.database.vectordb.chunkers.chunk_strategy import ChunkStrategy
|
||||
from cognee.database.vectordb.chunkers.chunk_strategy import ChunkStrategy
|
||||
from langchain.text_splitter import RecursiveCharacterTextSplitter
|
||||
|
||||
|
||||
|
|
|
|||
|
|
@ -3,8 +3,8 @@ import fitz
|
|||
import os
|
||||
import sys
|
||||
|
||||
from cognitive_architecture.database.vectordb.chunkers.chunkers import chunk_data
|
||||
from cognitive_architecture.shared.language_processing import (
|
||||
from cognee.database.vectordb.chunkers.chunkers import chunk_data
|
||||
from cognee.shared.language_processing import (
|
||||
translate_text,
|
||||
detect_language,
|
||||
)
|
||||
|
|
@ -148,7 +148,7 @@ async def _document_loader(observation: str, loader_settings: dict):
|
|||
# file_content += page.get_text()
|
||||
# pages = chunk_data(chunk_strategy=loader_strategy, source_data=file_content, chunk_size=chunk_size,
|
||||
# chunk_overlap=chunk_overlap)
|
||||
# from cognitive_architecture.shared.language_processing import translate_text,detect_language
|
||||
# from cognee.shared.language_processing import translate_text,detect_language
|
||||
#
|
||||
# if detect_language(pages) != "en":
|
||||
# logging.info("Current Directory 3")
|
||||
|
|
@ -169,7 +169,7 @@ async def _document_loader(observation: str, loader_settings: dict):
|
|||
# pages = chunk_data(chunk_strategy=loader_strategy, source_data=file_content, chunk_size=chunk_size,
|
||||
# chunk_overlap=chunk_overlap)
|
||||
#
|
||||
# from cognitive_architecture.shared.language_processing import translate_text, detect_language
|
||||
# from cognee.shared.language_processing import translate_text, detect_language
|
||||
#
|
||||
# if detect_language(pages) != "en":
|
||||
# logging.info("Current Directory 3")
|
||||
|
|
@ -196,7 +196,7 @@ async def _document_loader(observation: str, loader_settings: dict):
|
|||
# pages = chunk_data(chunk_strategy=loader_strategy, source_data=str(documents), chunk_size=chunk_size,
|
||||
# chunk_overlap=chunk_overlap)
|
||||
# logging.info("Documents: %s", documents)
|
||||
# from cognitive_architecture.shared.language_processing import translate_text, detect_language
|
||||
# from cognee.shared.language_processing import translate_text, detect_language
|
||||
#
|
||||
# if detect_language(pages) != "en":
|
||||
# logging.info("Current Directory 3")
|
||||
|
|
|
|||
|
|
@ -3,7 +3,7 @@ import logging
|
|||
|
||||
from langchain.text_splitter import RecursiveCharacterTextSplitter
|
||||
from marshmallow import Schema, fields
|
||||
from cognitive_architecture.database.vectordb.loaders.loaders import _document_loader
|
||||
from cognee.database.vectordb.loaders.loaders import _document_loader
|
||||
|
||||
# Add the parent directory to sys.path
|
||||
|
||||
|
|
@ -186,7 +186,7 @@ class WeaviateVectorDB(VectorDB):
|
|||
)
|
||||
else:
|
||||
chunk_count = 0
|
||||
from cognitive_architecture.database.vectordb.chunkers.chunkers import (
|
||||
from cognee.database.vectordb.chunkers.chunkers import (
|
||||
chunk_data,
|
||||
)
|
||||
|
||||
|
|
|
|||
|
|
@ -1,4 +1,4 @@
|
|||
from cognitive_architecture.config import Config
|
||||
from cognee.config import Config
|
||||
from .databases.relational import SqliteEngine, DatabaseEngine
|
||||
|
||||
config = Config()
|
||||
|
|
|
|||
|
|
@ -1,8 +1,8 @@
|
|||
import logging
|
||||
from . import Dataset, Data
|
||||
from cognitive_architecture.infrastructure import infrastructure_config
|
||||
from cognitive_architecture.infrastructure.databases.relational import DatabaseEngine
|
||||
from cognitive_architecture.infrastructure.files import remove_file_from_storage
|
||||
from cognee.infrastructure import infrastructure_config
|
||||
from cognee.infrastructure.databases.relational import DatabaseEngine
|
||||
from cognee.infrastructure.files import remove_file_from_storage
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
|
|
|
|||
|
|
@ -2,7 +2,7 @@ from typing import List
|
|||
from datetime import datetime, timezone
|
||||
from sqlalchemy.orm import relationship, MappedColumn, Mapped
|
||||
from sqlalchemy import Column, String, DateTime, UUID, Text, JSON
|
||||
from cognitive_architecture.infrastructure.databases.relational import ModelBase
|
||||
from cognee.infrastructure.databases.relational import ModelBase
|
||||
from .DatasetData import DatasetData
|
||||
|
||||
class Data(ModelBase):
|
||||
|
|
|
|||
|
|
@ -2,7 +2,7 @@ from typing import List
|
|||
from datetime import datetime, timezone
|
||||
from sqlalchemy.orm import relationship, Mapped
|
||||
from sqlalchemy import Column, Text, DateTime, UUID
|
||||
from cognitive_architecture.infrastructure.databases.relational import ModelBase
|
||||
from cognee.infrastructure.databases.relational import ModelBase
|
||||
from .DatasetData import DatasetData
|
||||
|
||||
class Dataset(ModelBase):
|
||||
|
|
|
|||
|
|
@ -1,7 +1,7 @@
|
|||
from uuid import uuid4
|
||||
from datetime import datetime, timezone
|
||||
from sqlalchemy import Column, DateTime, UUID, ForeignKey
|
||||
from cognitive_architecture.infrastructure.databases.relational import ModelBase
|
||||
from cognee.infrastructure.databases.relational import ModelBase
|
||||
|
||||
class DatasetData(ModelBase):
|
||||
__tablename__ = "dataset_data"
|
||||
|
|
|
|||
|
|
@ -1,7 +1,7 @@
|
|||
"""Factory function to get the appropriate graph client based on the graph type."""
|
||||
from cognitive_architecture.config import Config
|
||||
from cognitive_architecture.root_dir import get_absolute_path
|
||||
from cognitive_architecture.shared.data_models import GraphDBType
|
||||
from cognee.config import Config
|
||||
from cognee.root_dir import get_absolute_path
|
||||
from cognee.shared.data_models import GraphDBType
|
||||
from .graph_db_interface import GraphDBInterface
|
||||
from .networkx.adapter import NetworXAdapter
|
||||
# from .neo4j.adapter import Neo4jAdapter
|
||||
|
|
|
|||
|
|
@ -8,7 +8,7 @@ from typing import Optional, Dict, Any
|
|||
import aiofiles.os
|
||||
import aiofiles
|
||||
import networkx as nx
|
||||
from cognitive_architecture.infrastructure.databases.graph.graph_db_interface import GraphDBInterface
|
||||
from cognee.infrastructure.databases.graph.graph_db_interface import GraphDBInterface
|
||||
import logging
|
||||
|
||||
class NetworXAdapter(GraphDBInterface):
|
||||
|
|
|
|||
|
|
@ -1,5 +1,5 @@
|
|||
import duckdb
|
||||
from cognitive_architecture.root_dir import get_absolute_path
|
||||
from cognee.root_dir import get_absolute_path
|
||||
|
||||
class DuckDBAdapter():
|
||||
def __init__(self):
|
||||
|
|
|
|||
|
|
@ -2,9 +2,9 @@ import uuid
|
|||
from pathlib import Path
|
||||
from sqlalchemy import select
|
||||
from sqlalchemy.ext.asyncio import create_async_engine, async_sessionmaker, AsyncSession
|
||||
from cognitive_architecture.config import Config
|
||||
from cognee.config import Config
|
||||
# from ..relational_db_interface import RelationalDBInterface
|
||||
from cognitive_architecture.database.relationaldb.models.memory import MemoryModel
|
||||
from cognee.database.relationaldb.models.memory import MemoryModel
|
||||
|
||||
config = Config()
|
||||
config.load()
|
||||
|
|
|
|||
|
|
@ -2,7 +2,7 @@
|
|||
# from sqlalchemy.orm import relationship
|
||||
# # from sqlalchemy.orm import DeclarativeBase
|
||||
# from sqlalchemy import Column, String, DateTime, ForeignKey
|
||||
# from cognitive_architecture.database.relationaldb.database import Base
|
||||
# from cognee.database.relationaldb.database import Base
|
||||
|
||||
|
||||
# class MemoryModel(Base):
|
||||
|
|
|
|||
|
|
@ -4,7 +4,7 @@ from typing import Callable
|
|||
from sqlalchemy.inspection import inspect
|
||||
from sqlalchemy.ext.asyncio import create_async_engine, async_sessionmaker, AsyncEngine, AsyncSession, async_scoped_session
|
||||
from sqlalchemy.future import select
|
||||
from cognitive_architecture.infrastructure.files.storage.LocalStorage import LocalStorage
|
||||
from cognee.infrastructure.files.storage.LocalStorage import LocalStorage
|
||||
from ..DatabaseEngine import DatabaseEngine
|
||||
from ..ModelBase import ModelBase
|
||||
from ..utils import with_rollback
|
||||
|
|
|
|||
|
|
@ -1,4 +1,4 @@
|
|||
from cognitive_architecture.config import Config
|
||||
from cognee.config import Config
|
||||
from .qdrant import QDrantAdapter
|
||||
|
||||
config = Config()
|
||||
|
|
|
|||
|
|
@ -3,7 +3,7 @@ from langchain.embeddings.openai import OpenAIEmbeddings
|
|||
from langchain.retrievers import WeaviateHybridSearchRetriever, ParentDocumentRetriever
|
||||
from databases.vector.vector_db_interface import VectorDBInterface
|
||||
# from langchain.text_splitter import RecursiveCharacterTextSplitter
|
||||
from cognitive_architecture.database.vectordb.loaders.loaders import _document_loader
|
||||
from cognee.database.vectordb.loaders.loaders import _document_loader
|
||||
|
||||
class WeaviateVectorDB(VectorDBInterface):
|
||||
def __init__(self, *args, **kwargs):
|
||||
|
|
@ -136,7 +136,7 @@ class WeaviateVectorDB(VectorDBInterface):
|
|||
)
|
||||
else:
|
||||
chunk_count = 0
|
||||
from cognitive_architecture.database.vectordb.chunkers.chunkers import (
|
||||
from cognee.database.vectordb.chunkers.chunkers import (
|
||||
chunk_data,
|
||||
)
|
||||
|
||||
|
|
|
|||
|
|
@ -1,5 +1,5 @@
|
|||
from typing import BinaryIO
|
||||
from cognitive_architecture.root_dir import get_absolute_path
|
||||
from cognee.root_dir import get_absolute_path
|
||||
from .storage.StorageManager import StorageManager
|
||||
from .storage.LocalStorage import LocalStorage
|
||||
|
||||
|
|
|
|||
|
|
@ -1,4 +1,4 @@
|
|||
from cognitive_architecture.root_dir import get_absolute_path
|
||||
from cognee.root_dir import get_absolute_path
|
||||
from .storage.StorageManager import StorageManager
|
||||
from .storage.LocalStorage import LocalStorage
|
||||
|
||||
|
|
|
|||
|
|
@ -1,5 +1,5 @@
|
|||
"""Get the LLM client."""
|
||||
from cognitive_architecture.config import Config
|
||||
from cognee.config import Config
|
||||
from .openai.adapter import OpenAIAdapter
|
||||
|
||||
config = Config()
|
||||
|
|
|
|||
|
|
@ -6,7 +6,7 @@ import instructor
|
|||
from openai import AsyncOpenAI
|
||||
from pydantic import BaseModel
|
||||
from tenacity import retry, stop_after_attempt
|
||||
from cognitive_architecture.utils import read_query_prompt
|
||||
from cognee.utils import read_query_prompt
|
||||
from ..llm_interface import LLMInterface
|
||||
|
||||
class OpenAIAdapter(LLMInterface):
|
||||
|
|
|
|||
|
|
@ -3,8 +3,8 @@ import os
|
|||
import instructor
|
||||
from openai import OpenAI
|
||||
import logging
|
||||
from cognitive_architecture.shared.data_models import KnowledgeGraph, MemorySummary
|
||||
from cognitive_architecture.config import Config
|
||||
from cognee.shared.data_models import KnowledgeGraph, MemorySummary
|
||||
from cognee.config import Config
|
||||
|
||||
|
||||
|
||||
|
|
@ -36,7 +36,7 @@ def generate_graph(input) -> KnowledgeGraph:
|
|||
model = "gpt-4-1106-preview"
|
||||
user_prompt = f"Use the given format to extract information from the following input: {input}."
|
||||
system_prompt = read_query_prompt(
|
||||
"cognitive_architecture/llm/prompts/generate_graph_prompt.txt"
|
||||
"cognee/llm/prompts/generate_graph_prompt.txt"
|
||||
)
|
||||
|
||||
out = aclient.chat.completions.create(
|
||||
|
|
@ -87,7 +87,7 @@ async def generate_summary(input) -> MemorySummary:
|
|||
def user_query_to_edges_and_nodes(input: str) -> KnowledgeGraph:
|
||||
"""Generate a knowledge graph from a user query."""
|
||||
system_prompt = read_query_prompt(
|
||||
"cognitive_architecture/llm/prompts/generate_graph_prompt.txt"
|
||||
"cognee/llm/prompts/generate_graph_prompt.txt"
|
||||
)
|
||||
return aclient.chat.completions.create(
|
||||
model=config.model,
|
||||
|
|
|
|||
|
|
@ -1,7 +1,7 @@
|
|||
from datetime import datetime
|
||||
from sqlalchemy.orm import Mapped, MappedColumn
|
||||
from sqlalchemy import Column, String, DateTime, ForeignKey, Enum, UUID, JSON
|
||||
from cognitive_architecture.infrastructure.databases.relational import ModelBase
|
||||
from cognee.infrastructure.databases.relational import ModelBase
|
||||
|
||||
class OperationType(Enum):
|
||||
MERGE_DATA = "MERGE_DATA"
|
||||
|
|
|
|||
|
|
@ -1,5 +1,5 @@
|
|||
""" Here we update semantic graph with content that classifier produced"""
|
||||
from cognitive_architecture.infrastructure.databases.graph.get_graph_client import get_graph_client, GraphDBType
|
||||
from cognee.infrastructure.databases.graph.get_graph_client import get_graph_client, GraphDBType
|
||||
|
||||
|
||||
async def add_classification_nodes(document_id, classification_data):
|
||||
|
|
|
|||
|
|
@ -1,5 +1,5 @@
|
|||
from cognitive_architecture.infrastructure.databases.graph.get_graph_client import get_graph_client
|
||||
from cognitive_architecture.shared.data_models import GraphDBType, Document, DocumentType, Category, Relationship
|
||||
from cognee.infrastructure.databases.graph.get_graph_client import get_graph_client
|
||||
from cognee.shared.data_models import GraphDBType, Document, DocumentType, Category, Relationship
|
||||
from .create import add_node_and_edge
|
||||
|
||||
def create_category(category_name: str):
|
||||
|
|
|
|||
|
|
@ -1,6 +1,6 @@
|
|||
from networkx import Graph
|
||||
from cognitive_architecture.infrastructure.databases.graph.get_graph_client import get_graph_client
|
||||
from cognitive_architecture.shared.data_models import GraphDBType
|
||||
from cognee.infrastructure.databases.graph.get_graph_client import get_graph_client
|
||||
from cognee.shared.data_models import GraphDBType
|
||||
|
||||
|
||||
async def extract_node_descriptions(data):
|
||||
|
|
|
|||
|
|
@ -2,7 +2,7 @@
|
|||
import uuid
|
||||
import json
|
||||
from datetime import datetime
|
||||
from cognitive_architecture.infrastructure.databases.graph.get_graph_client import get_graph_client, GraphDBType
|
||||
from cognee.infrastructure.databases.graph.get_graph_client import get_graph_client, GraphDBType
|
||||
|
||||
|
||||
async def add_propositions(
|
||||
|
|
|
|||
|
|
@ -1,8 +1,8 @@
|
|||
""" This module is responsible for creating a semantic graph """
|
||||
from typing import Optional, Any
|
||||
from pydantic import BaseModel
|
||||
from cognitive_architecture.infrastructure.databases.graph.get_graph_client import get_graph_client
|
||||
from cognitive_architecture.shared.data_models import GraphDBType
|
||||
from cognee.infrastructure.databases.graph.get_graph_client import get_graph_client
|
||||
from cognee.shared.data_models import GraphDBType
|
||||
|
||||
|
||||
async def generate_node_id(instance: BaseModel) -> str:
|
||||
|
|
|
|||
|
|
@ -1,6 +1,6 @@
|
|||
from datetime import datetime
|
||||
from cognitive_architecture.shared.data_models import DefaultGraphModel, Relationship, UserProperties, UserLocation
|
||||
from cognitive_architecture.modules.cognify.graph.create import create_semantic_graph
|
||||
from cognee.shared.data_models import DefaultGraphModel, Relationship, UserProperties, UserLocation
|
||||
from cognee.modules.cognify.graph.create import create_semantic_graph
|
||||
|
||||
async def initialize_graph(root_id: str):
|
||||
graph = DefaultGraphModel(
|
||||
|
|
|
|||
|
|
@ -1,8 +1,8 @@
|
|||
""" This module contains the code to classify content into categories using the LLM API. """
|
||||
from typing import Type, List
|
||||
from pydantic import BaseModel
|
||||
from cognitive_architecture.infrastructure.llm.get_llm_client import get_llm_client
|
||||
from cognitive_architecture.utils import read_query_prompt
|
||||
from cognee.infrastructure.llm.get_llm_client import get_llm_client
|
||||
from cognee.utils import read_query_prompt
|
||||
|
||||
async def classify_into_categories(text_input: str, system_prompt_path: str, response_model: Type[BaseModel]):
|
||||
llm_client = get_llm_client()
|
||||
|
|
|
|||
|
|
@ -1,8 +1,8 @@
|
|||
""" This module contains the code to classify content into categories using the LLM API. """
|
||||
from typing import Type
|
||||
from pydantic import BaseModel
|
||||
from cognitive_architecture.infrastructure.llm.get_llm_client import get_llm_client
|
||||
from cognitive_architecture.utils import async_render_template
|
||||
from cognee.infrastructure.llm.get_llm_client import get_llm_client
|
||||
from cognee.utils import async_render_template
|
||||
|
||||
async def content_to_cog_layers(filename: str, context, response_model: Type[BaseModel]):
|
||||
llm_client = get_llm_client()
|
||||
|
|
|
|||
|
|
@ -2,9 +2,9 @@
|
|||
import json
|
||||
from typing import Type
|
||||
from pydantic import BaseModel
|
||||
from cognitive_architecture.infrastructure.llm.get_llm_client import get_llm_client
|
||||
from cognitive_architecture.shared.data_models import KnowledgeGraph
|
||||
from cognitive_architecture.utils import async_render_template
|
||||
from cognee.infrastructure.llm.get_llm_client import get_llm_client
|
||||
from cognee.shared.data_models import KnowledgeGraph
|
||||
from cognee.utils import async_render_template
|
||||
|
||||
async def generate_graph(text_input: str, filename: str, context, response_model: Type[BaseModel]):
|
||||
llm_client = get_llm_client()
|
||||
|
|
|
|||
|
|
@ -1,8 +1,8 @@
|
|||
import asyncio
|
||||
|
||||
from qdrant_client import models
|
||||
from cognitive_architecture.infrastructure.llm.get_llm_client import get_llm_client
|
||||
from cognitive_architecture.infrastructure.databases.vector import get_vector_database
|
||||
from cognee.infrastructure.llm.get_llm_client import get_llm_client
|
||||
from cognee.infrastructure.databases.vector import get_vector_database
|
||||
|
||||
async def get_embeddings(texts:list):
|
||||
""" Get embeddings for a list of texts"""
|
||||
|
|
|
|||
|
|
@ -1,5 +1,5 @@
|
|||
from typing import BinaryIO
|
||||
from cognitive_architecture.infrastructure.files import get_file_metadata, FileMetadata
|
||||
from cognee.infrastructure.files import get_file_metadata, FileMetadata
|
||||
from .IngestionData import IngestionData
|
||||
|
||||
def create_binary_data(data: BinaryIO):
|
||||
|
|
|
|||
|
|
@ -1,7 +1,7 @@
|
|||
import asyncio
|
||||
from uuid import UUID, uuid4
|
||||
from cognitive_architecture.infrastructure.files import add_file_to_storage
|
||||
from cognitive_architecture.infrastructure.data import add_data_to_dataset, Data, Dataset
|
||||
from cognee.infrastructure.files import add_file_to_storage
|
||||
from cognee.infrastructure.data import add_data_to_dataset, Data, Dataset
|
||||
from .data_types import IngestionData
|
||||
|
||||
async def save(dataset_id: UUID, dataset_name: str, data_id: UUID, data: IngestionData):
|
||||
|
|
|
|||
|
|
@ -1,5 +1,5 @@
|
|||
from cognitive_architecture.infrastructure.databases.vector.qdrant.adapter import CollectionConfig
|
||||
from cognitive_architecture.infrastructure.databases.vector.get_vector_database import get_vector_database
|
||||
from cognee.infrastructure.databases.vector.qdrant.adapter import CollectionConfig
|
||||
from cognee.infrastructure.databases.vector.get_vector_database import get_vector_database
|
||||
|
||||
async def create_vector_memory(memory_name: str, collection_config: CollectionConfig):
|
||||
vector_db = get_vector_database()
|
||||
|
|
|
|||
|
|
@ -1,5 +1,5 @@
|
|||
""" Fetches the context of a given node in the graph"""
|
||||
from cognitive_architecture.infrastructure.databases.graph.get_graph_client import get_graph_client
|
||||
from cognee.infrastructure.databases.graph.get_graph_client import get_graph_client
|
||||
async def search_neighbour(CONNECTED_GRAPH, id):
|
||||
relevant_context = []
|
||||
for n,attr in CONNECTED_GRAPH.nodes(data=True):
|
||||
|
|
@ -18,7 +18,7 @@ async def search_neighbour(CONNECTED_GRAPH, id):
|
|||
if __name__ == '__main__':
|
||||
import asyncio
|
||||
async def main():
|
||||
from cognitive_architecture.shared.data_models import GraphDBType
|
||||
from cognee.shared.data_models import GraphDBType
|
||||
|
||||
graph_client = get_graph_client(GraphDBType.NETWORKX)
|
||||
graph = await graph_client.graph
|
||||
|
|
|
|||
|
|
@ -1,6 +1,6 @@
|
|||
|
||||
from cognitive_architecture.infrastructure.llm.get_llm_client import get_llm_client
|
||||
from cognitive_architecture.modules.cognify.graph.add_node_connections import extract_node_descriptions
|
||||
from cognee.infrastructure.llm.get_llm_client import get_llm_client
|
||||
from cognee.modules.cognify.graph.add_node_connections import extract_node_descriptions
|
||||
|
||||
|
||||
async def search_similarity(query:str ,graph):
|
||||
|
|
@ -15,7 +15,7 @@ async def search_similarity(query:str ,graph):
|
|||
query = await client.async_get_embedding_with_backoff(query)
|
||||
# print(query)
|
||||
for id in unique_layer_uuids:
|
||||
from cognitive_architecture.infrastructure.databases.vector.get_vector_database import get_vector_database
|
||||
from cognee.infrastructure.databases.vector.get_vector_database import get_vector_database
|
||||
vector_client = get_vector_database()
|
||||
|
||||
print(query)
|
||||
|
|
|
|||
|
|
@ -1,8 +1,8 @@
|
|||
import uuid
|
||||
from typing import List
|
||||
from qdrant_client.models import PointStruct
|
||||
from cognitive_architecture.infrastructure.databases.vector.get_vector_database import get_vector_database
|
||||
from cognitive_architecture.infrastructure.llm.openai.openai_tools import async_get_embedding_with_backoff
|
||||
from cognee.infrastructure.databases.vector.get_vector_database import get_vector_database
|
||||
from cognee.infrastructure.llm.openai.openai_tools import async_get_embedding_with_backoff
|
||||
|
||||
async def create_information_points(memory_name: str, payload: List[str]):
|
||||
vector_db = get_vector_database()
|
||||
|
|
|
|||
|
|
@ -1,4 +1,4 @@
|
|||
from cognitive_architecture.infrastructure.databases.relational.get_database import get_database
|
||||
from cognee.infrastructure.databases.relational.get_database import get_database
|
||||
|
||||
async def is_existing_memory(memory_name: str):
|
||||
memory = await (get_database().get_memory_by_name(memory_name))
|
||||
|
|
|
|||
|
|
@ -1,4 +1,4 @@
|
|||
from cognitive_architecture.infrastructure.databases.relational.get_database import get_database
|
||||
from cognee.infrastructure.databases.relational.get_database import get_database
|
||||
|
||||
def register_memory_for_user(user_id: str, memory_name: str):
|
||||
return get_database().add_memory(user_id, memory_name)
|
||||
|
|
|
|||
|
|
@ -5,11 +5,11 @@ logger = logging.getLogger(__name__)
|
|||
|
||||
async def main():
|
||||
"""Runs as a part of startup docker scripts to create the database and tables."""
|
||||
from cognitive_architecture.config import Config
|
||||
from cognee.config import Config
|
||||
config = Config()
|
||||
config.load()
|
||||
|
||||
from cognitive_architecture.database.database_manager import DatabaseManager
|
||||
from cognee.database.database_manager import DatabaseManager
|
||||
|
||||
db_manager = DatabaseManager()
|
||||
database_name = config.db_name
|
||||
|
|
|
|||
|
|
@ -12,11 +12,11 @@ from sqlalchemy import or_
|
|||
from sqlalchemy.future import select
|
||||
from sqlalchemy.orm import contains_eager
|
||||
from sqlalchemy.ext.asyncio import AsyncSession
|
||||
from cognitive_architecture.database.relationaldb.models.docs import DocsModel
|
||||
from cognitive_architecture.database.relationaldb.models.memory import MemoryModel
|
||||
from cognitive_architecture.database.relationaldb.models.operation import Operation
|
||||
from cognee.database.relationaldb.models.docs import DocsModel
|
||||
from cognee.database.relationaldb.models.memory import MemoryModel
|
||||
from cognee.database.relationaldb.models.operation import Operation
|
||||
|
||||
from cognitive_architecture.config import Config
|
||||
from cognee.config import Config
|
||||
|
||||
config = Config()
|
||||
config.load()
|
||||
|
|
@ -290,7 +290,7 @@ async def read_query_prompt(filename: str) -> str:
|
|||
script_directory = Path(__file__).parent
|
||||
|
||||
# Set the base directory relative to the script's directory
|
||||
base_directory = script_directory.parent / "cognitive_architecture/infrastructure/llm/prompts"
|
||||
base_directory = script_directory.parent / "cognee/infrastructure/llm/prompts"
|
||||
|
||||
# Construct the full file path
|
||||
file_path = base_directory / filename
|
||||
|
|
@ -326,7 +326,7 @@ async def async_render_template(filename: str, context: dict) -> str:
|
|||
script_directory = Path(__file__).parent
|
||||
|
||||
# Set the base directory relative to the script's directory
|
||||
base_directory = script_directory.parent / "cognitive_architecture/infrastructure/llm/prompts"
|
||||
base_directory = script_directory.parent / "cognee/infrastructure/llm/prompts"
|
||||
|
||||
|
||||
# Construct the full file path
|
||||
|
|
|
|||
|
|
@ -3,15 +3,15 @@ import logging
|
|||
logging.basicConfig(level=logging.INFO)
|
||||
|
||||
from sqlalchemy.future import select
|
||||
from cognitive_architecture.database.relationaldb.models.user import User
|
||||
from cognitive_architecture.database.relationaldb.models.memory import MemoryModel
|
||||
from cognee.database.relationaldb.models.user import User
|
||||
from cognee.database.relationaldb.models.memory import MemoryModel
|
||||
import ast
|
||||
import tracemalloc
|
||||
from cognitive_architecture.database.relationaldb.database_crud import add_entity
|
||||
from cognee.database.relationaldb.database_crud import add_entity
|
||||
tracemalloc.start()
|
||||
import uuid
|
||||
from cognitive_architecture.database.vectordb.basevectordb import BaseMemory
|
||||
from cognitive_architecture.config import Config
|
||||
from cognee.database.vectordb.basevectordb import BaseMemory
|
||||
from cognee.config import Config
|
||||
|
||||
globalConfig = Config()
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ echo "Environment: $ENVIRONMENT"
|
|||
if [ "$ENVIRONMENT" != "local" ]; then
|
||||
echo "Running fetch_secret.py"
|
||||
|
||||
PYTHONPATH=. python cognitive_architecture/fetch_secret.py
|
||||
PYTHONPATH=. python cognee/fetch_secret.py
|
||||
|
||||
if [ $? -ne 0 ]; then
|
||||
echo "Error: fetch_secret.py failed"
|
||||
|
|
@ -19,7 +19,7 @@ fi
|
|||
|
||||
echo "Creating database..."
|
||||
|
||||
PYTHONPATH=. python cognitive_architecture/setup_database.py
|
||||
PYTHONPATH=. python cognee/setup_database.py
|
||||
if [ $? -ne 0 ]; then
|
||||
echo "Error: setup_database.py failed"
|
||||
exit 1
|
||||
|
|
|
|||
46
main.py
46
main.py
|
|
@ -2,53 +2,53 @@ from typing import Optional, List
|
|||
|
||||
from neo4j.exceptions import Neo4jError
|
||||
from pydantic import BaseModel, Field
|
||||
from cognitive_architecture.database.graphdb.graph import Neo4jGraphDB
|
||||
from cognitive_architecture.database.relationaldb.models.memory import MemoryModel
|
||||
from cognee.database.graphdb.graph import Neo4jGraphDB
|
||||
from cognee.database.relationaldb.models.memory import MemoryModel
|
||||
import os
|
||||
from cognitive_architecture.database.relationaldb.database_crud import (
|
||||
from cognee.database.relationaldb.database_crud import (
|
||||
session_scope,
|
||||
update_entity_graph_summary,
|
||||
)
|
||||
from cognitive_architecture.database.relationaldb.database import AsyncSessionLocal
|
||||
from cognitive_architecture.utils import generate_letter_uuid
|
||||
from cognee.database.relationaldb.database import AsyncSessionLocal
|
||||
from cognee.utils import generate_letter_uuid
|
||||
import instructor
|
||||
from openai import OpenAI
|
||||
from cognitive_architecture.vectorstore_manager import Memory
|
||||
from cognitive_architecture.database.relationaldb.database_crud import fetch_job_id
|
||||
from cognee.vectorstore_manager import Memory
|
||||
from cognee.database.relationaldb.database_crud import fetch_job_id
|
||||
import uuid
|
||||
from cognitive_architecture.database.relationaldb.models.sessions import Session
|
||||
from cognitive_architecture.database.relationaldb.models.operation import Operation
|
||||
from cognitive_architecture.database.relationaldb.database_crud import (
|
||||
from cognee.database.relationaldb.models.sessions import Session
|
||||
from cognee.database.relationaldb.models.operation import Operation
|
||||
from cognee.database.relationaldb.database_crud import (
|
||||
session_scope,
|
||||
add_entity,
|
||||
update_entity,
|
||||
fetch_job_id,
|
||||
)
|
||||
from cognitive_architecture.database.relationaldb.models.metadatas import MetaDatas
|
||||
from cognitive_architecture.database.relationaldb.models.docs import DocsModel
|
||||
from cognitive_architecture.database.relationaldb.models.memory import MemoryModel
|
||||
from cognitive_architecture.database.relationaldb.models.user import User
|
||||
from cognitive_architecture.classifiers.classify_summary import classify_summary
|
||||
from cognitive_architecture.classifiers.classify_documents import classify_documents
|
||||
from cognitive_architecture.classifiers.classify_user_query import classify_user_query
|
||||
from cognitive_architecture.classifiers.classify_user_input import classify_user_input
|
||||
from cognee.database.relationaldb.models.metadatas import MetaDatas
|
||||
from cognee.database.relationaldb.models.docs import DocsModel
|
||||
from cognee.database.relationaldb.models.memory import MemoryModel
|
||||
from cognee.database.relationaldb.models.user import User
|
||||
from cognee.classifiers.classify_summary import classify_summary
|
||||
from cognee.classifiers.classify_documents import classify_documents
|
||||
from cognee.classifiers.classify_user_query import classify_user_query
|
||||
from cognee.classifiers.classify_user_input import classify_user_input
|
||||
|
||||
aclient = instructor.patch(OpenAI())
|
||||
DEFAULT_PRESET = "promethai_chat"
|
||||
preset_options = [DEFAULT_PRESET]
|
||||
PROMETHAI_DIR = os.path.join(os.path.expanduser("~"), ".")
|
||||
OPENAI_API_KEY = os.getenv("OPENAI_API_KEY", "")
|
||||
from cognitive_architecture.config import Config
|
||||
from cognee.config import Config
|
||||
|
||||
config = Config()
|
||||
config.load()
|
||||
|
||||
from cognitive_architecture.utils import get_document_names
|
||||
from cognee.utils import get_document_names
|
||||
from sqlalchemy.orm import selectinload, joinedload, contains_eager
|
||||
import logging
|
||||
from sqlalchemy.ext.asyncio import AsyncSession
|
||||
from sqlalchemy.future import select
|
||||
from cognitive_architecture.utils import (
|
||||
from cognee.utils import (
|
||||
get_document_names,
|
||||
generate_letter_uuid,
|
||||
get_memory_name_by_doc_id,
|
||||
|
|
@ -56,7 +56,7 @@ from cognitive_architecture.utils import (
|
|||
get_vectordb_namespace,
|
||||
get_vectordb_document_name,
|
||||
)
|
||||
from cognitive_architecture.shared.language_processing import (
|
||||
from cognee.shared.language_processing import (
|
||||
translate_text,
|
||||
detect_language,
|
||||
)
|
||||
|
|
@ -152,7 +152,7 @@ async def load_documents_to_vectorstore(
|
|||
# except:
|
||||
# document_names = document_source
|
||||
for doc in document_names:
|
||||
from cognitive_architecture.shared.language_processing import (
|
||||
from cognee.shared.language_processing import (
|
||||
translate_text,
|
||||
detect_language,
|
||||
)
|
||||
|
|
|
|||
6
poetry.lock
generated
6
poetry.lock
generated
|
|
@ -4096,7 +4096,6 @@ description = "Nvidia JIT LTO Library"
|
|||
optional = false
|
||||
python-versions = ">=3"
|
||||
files = [
|
||||
{file = "nvidia_nvjitlink_cu12-12.4.99-py3-none-manylinux2014_aarch64.whl", hash = "sha256:75d6498c96d9adb9435f2bbdbddb479805ddfb97b5c1b32395c694185c20ca57"},
|
||||
{file = "nvidia_nvjitlink_cu12-12.4.99-py3-none-manylinux2014_x86_64.whl", hash = "sha256:c6428836d20fe7e327191c175791d38570e10762edc588fb46749217cd444c74"},
|
||||
{file = "nvidia_nvjitlink_cu12-12.4.99-py3-none-win_amd64.whl", hash = "sha256:991905ffa2144cb603d8ca7962d75c35334ae82bf92820b6ba78157277da1ad2"},
|
||||
]
|
||||
|
|
@ -5701,6 +5700,7 @@ files = [
|
|||
{file = "PyYAML-6.0.1-cp311-cp311-win_amd64.whl", hash = "sha256:bf07ee2fef7014951eeb99f56f39c9bb4af143d8aa3c21b1677805985307da34"},
|
||||
{file = "PyYAML-6.0.1-cp312-cp312-macosx_10_9_x86_64.whl", hash = "sha256:855fb52b0dc35af121542a76b9a84f8d1cd886ea97c84703eaa6d88e37a2ad28"},
|
||||
{file = "PyYAML-6.0.1-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:40df9b996c2b73138957fe23a16a4f0ba614f4c0efce1e9406a184b6d07fa3a9"},
|
||||
{file = "PyYAML-6.0.1-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a08c6f0fe150303c1c6b71ebcd7213c2858041a7e01975da3a99aed1e7a378ef"},
|
||||
{file = "PyYAML-6.0.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6c22bec3fbe2524cde73d7ada88f6566758a8f7227bfbf93a408a9d86bcc12a0"},
|
||||
{file = "PyYAML-6.0.1-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:8d4e9c88387b0f5c7d5f281e55304de64cf7f9c0021a3525bd3b1c542da3b0e4"},
|
||||
{file = "PyYAML-6.0.1-cp312-cp312-win32.whl", hash = "sha256:d483d2cdf104e7c9fa60c544d92981f12ad66a457afae824d146093b8c294c54"},
|
||||
|
|
@ -8057,5 +8057,5 @@ weaviate = []
|
|||
|
||||
[metadata]
|
||||
lock-version = "2.0"
|
||||
python-versions = "3.10.13"
|
||||
content-hash = "d9010f8123850150bf7a9cf1c2955a69c69ce28467bbe8b6ff9d7bf0c7f072e1"
|
||||
python-versions = "~3.10"
|
||||
content-hash = "48d8cbd3319b1370abddeb2ff6c32cf6be01bdaff150d466129a1667b8cce94f"
|
||||
|
|
|
|||
|
|
@ -17,7 +17,7 @@ classifiers = [
|
|||
"Operating System :: Microsoft :: Windows",]
|
||||
|
||||
[tool.poetry.dependencies]
|
||||
python = "3.10.13"
|
||||
python = "~3.10"
|
||||
langchain = "^0.0.338"
|
||||
openai = "1.12.0"
|
||||
python-dotenv = "1.0.1"
|
||||
|
|
|
|||
File diff suppressed because it is too large
Load diff
Loading…
Add table
Reference in a new issue