<!-- .github/pull_request_template.md --> This PR introduces wide vector and graph structure filtering capabilities. With these changes, the graph completion retriever and all retrievers that inherit from it will now filter relevant vector elements and subgraphs based on the query. This improvement significantly increases search speed for large graphs while maintaining—and in some cases slightly improving—accuracy. Changes in This PR: -Introduced new wide_search_top_k parameter: Controls the initial search space size -Added graph adapter level filtering method: Enables relevant subgraph filtering while maintaining backward compatibility. For community or custom graph adapters that don't implement this method, the system gracefully falls back to the original search behavior. -Updated modal dashboard and evaluation framework: Fixed compatibility issues. Added comprehensive unit tests: Introduced unit tests for brute_force_triplet_search (previously untested) and expanded the CogneeGraph test suite. Integration tests: Existing integration tests verify end-to-end search functionality (no changes required). Acceptance Criteria and Testing To verify the new search behavior, run search queries with different wide_search_top_k parameters while logging is enabled: None: Triggers a full graph search (default behavior) 1: Projects a minimal subgraph (demonstrates maximum filtering) Custom values: Test intermediate levels of filtering Internal Testing and results: Performance and accuracy benchmarks are available upon request. The implementation demonstrates measurable improvements in query latency for large graphs without sacrificing result quality. ## Type of Change <!-- Please check the relevant option --> - [ ] Bug fix (non-breaking change that fixes an issue) - [ ] New feature (non-breaking change that adds functionality) - [ ] Breaking change (fix or feature that would cause existing functionality to change) - [ ] Documentation update - [x] Code refactoring - [x] Performance improvement - [ ] Other (please specify): ## Screenshots/Videos (if applicable) None ## Pre-submission Checklist <!-- Please check all boxes that apply before submitting your PR --> - [x] **I have tested my changes thoroughly before submitting this PR** - [x] **This PR contains minimal changes necessary to address the issue/feature** - [x] My code follows the project's coding standards and style guidelines - [x] I have added tests that prove my fix is effective or that my feature works - [x] I have added necessary documentation (if applicable) - [x] All new and existing tests pass - [x] I have searched existing PRs to ensure this change hasn't been submitted already - [x] I have linked any relevant issues in the description - [x] My commits have clear and descriptive messages ## DCO Affirmation I affirm that all code in every commit of this pull request conforms to the terms of the Topoteretes Developer Certificate of Origin. --------- Co-authored-by: Pavel Zorin <pazonec@yandex.ru>
178 lines
6.4 KiB
Python
178 lines
6.4 KiB
Python
import asyncio
|
||
from typing import Optional, List, Type, Any
|
||
from cognee.modules.graph.cognee_graph.CogneeGraphElements import Edge
|
||
from cognee.shared.logging_utils import get_logger
|
||
from cognee.modules.retrieval.graph_completion_retriever import GraphCompletionRetriever
|
||
from cognee.modules.retrieval.utils.completion import generate_completion, summarize_text
|
||
from cognee.modules.retrieval.utils.session_cache import (
|
||
save_conversation_history,
|
||
get_conversation_history,
|
||
)
|
||
from cognee.context_global_variables import session_user
|
||
from cognee.infrastructure.databases.cache.config import CacheConfig
|
||
|
||
logger = get_logger()
|
||
|
||
|
||
class GraphCompletionContextExtensionRetriever(GraphCompletionRetriever):
|
||
"""
|
||
Handles graph context completion for question answering tasks, extending context based
|
||
on retrieved triplets.
|
||
|
||
Public methods:
|
||
- get_completion
|
||
|
||
Instance variables:
|
||
- user_prompt_path
|
||
- system_prompt_path
|
||
- top_k
|
||
- node_type
|
||
- node_name
|
||
"""
|
||
|
||
def __init__(
|
||
self,
|
||
user_prompt_path: str = "graph_context_for_question.txt",
|
||
system_prompt_path: str = "answer_simple_question.txt",
|
||
system_prompt: Optional[str] = None,
|
||
top_k: Optional[int] = 5,
|
||
node_type: Optional[Type] = None,
|
||
node_name: Optional[List[str]] = None,
|
||
save_interaction: bool = False,
|
||
wide_search_top_k: Optional[int] = 100,
|
||
triplet_distance_penalty: Optional[float] = 3.5,
|
||
):
|
||
super().__init__(
|
||
user_prompt_path=user_prompt_path,
|
||
system_prompt_path=system_prompt_path,
|
||
top_k=top_k,
|
||
node_type=node_type,
|
||
node_name=node_name,
|
||
save_interaction=save_interaction,
|
||
system_prompt=system_prompt,
|
||
wide_search_top_k=wide_search_top_k,
|
||
triplet_distance_penalty=triplet_distance_penalty,
|
||
)
|
||
|
||
async def get_completion(
|
||
self,
|
||
query: str,
|
||
context: Optional[List[Edge]] = None,
|
||
session_id: Optional[str] = None,
|
||
context_extension_rounds=4,
|
||
response_model: Type = str,
|
||
) -> List[Any]:
|
||
"""
|
||
Extends the context for a given query by retrieving related triplets and generating new
|
||
completions based on them.
|
||
|
||
The method runs for a specified number of rounds to enhance context until no new
|
||
triplets are found or the maximum rounds are reached. It retrieves triplet suggestions
|
||
based on a generated completion from previous iterations, logging the process of context
|
||
extension.
|
||
|
||
Parameters:
|
||
-----------
|
||
|
||
- query (str): The input query for which the completion is generated.
|
||
- context (Optional[Any]): The existing context to use for enhancing the query; if
|
||
None, it will be initialized from triplets generated for the query. (default None)
|
||
- session_id (Optional[str]): Optional session identifier for caching. If None,
|
||
defaults to 'default_session'. (default None)
|
||
- context_extension_rounds: The maximum number of rounds to extend the context with
|
||
new triplets before halting. (default 4)
|
||
- response_model (Type): The Pydantic model type for structured output. (default str)
|
||
|
||
Returns:
|
||
--------
|
||
|
||
- List[str]: A list containing the generated answer based on the query and the
|
||
extended context.
|
||
"""
|
||
triplets = context
|
||
|
||
if triplets is None:
|
||
triplets = await self.get_context(query)
|
||
|
||
context_text = await self.resolve_edges_to_text(triplets)
|
||
|
||
round_idx = 1
|
||
|
||
while round_idx <= context_extension_rounds:
|
||
prev_size = len(triplets)
|
||
|
||
logger.info(
|
||
f"Context extension: round {round_idx} - generating next graph locational query."
|
||
)
|
||
completion = await generate_completion(
|
||
query=query,
|
||
context=context_text,
|
||
user_prompt_path=self.user_prompt_path,
|
||
system_prompt_path=self.system_prompt_path,
|
||
system_prompt=self.system_prompt,
|
||
)
|
||
|
||
triplets += await self.get_context(completion)
|
||
triplets = list(set(triplets))
|
||
context_text = await self.resolve_edges_to_text(triplets)
|
||
|
||
num_triplets = len(triplets)
|
||
|
||
if num_triplets == prev_size:
|
||
logger.info(
|
||
f"Context extension: round {round_idx} – no new triplets found; stopping early."
|
||
)
|
||
break
|
||
|
||
logger.info(
|
||
f"Context extension: round {round_idx} - "
|
||
f"number of unique retrieved triplets: {num_triplets}"
|
||
)
|
||
|
||
round_idx += 1
|
||
|
||
# Check if we need to generate context summary for caching
|
||
cache_config = CacheConfig()
|
||
user = session_user.get()
|
||
user_id = getattr(user, "id", None)
|
||
session_save = user_id and cache_config.caching
|
||
|
||
if session_save:
|
||
conversation_history = await get_conversation_history(session_id=session_id)
|
||
|
||
context_summary, completion = await asyncio.gather(
|
||
summarize_text(context_text),
|
||
generate_completion(
|
||
query=query,
|
||
context=context_text,
|
||
user_prompt_path=self.user_prompt_path,
|
||
system_prompt_path=self.system_prompt_path,
|
||
system_prompt=self.system_prompt,
|
||
conversation_history=conversation_history,
|
||
response_model=response_model,
|
||
),
|
||
)
|
||
else:
|
||
completion = await generate_completion(
|
||
query=query,
|
||
context=context_text,
|
||
user_prompt_path=self.user_prompt_path,
|
||
system_prompt_path=self.system_prompt_path,
|
||
system_prompt=self.system_prompt,
|
||
response_model=response_model,
|
||
)
|
||
|
||
if self.save_interaction and context_text and triplets and completion:
|
||
await self.save_qa(
|
||
question=query, answer=completion, context=context_text, triplets=triplets
|
||
)
|
||
|
||
if session_save:
|
||
await save_conversation_history(
|
||
query=query,
|
||
context_summary=context_summary,
|
||
answer=completion,
|
||
session_id=session_id,
|
||
)
|
||
|
||
return [completion]
|