cognee/evals/old/comparative_eval
2025-08-17 12:19:34 +02:00
..
helpers move to gpt5 2025-08-17 12:19:34 +02:00
hotpot_50_corpus.json fix: Feature/cog 2648 evals update (#1221) 2025-08-08 20:23:09 +02:00
hotpot_50_qa_pairs.json fix: Feature/cog 2648 evals update (#1221) 2025-08-08 20:23:09 +02:00
README.md fix: Feature/cog 2648 evals update (#1221) 2025-08-08 20:23:09 +02:00

Comparative QA Benchmarks

Independent benchmarks for different QA/RAG systems using HotpotQA dataset.

Dataset Files

  • hotpot_50_corpus.json - 50 instances from HotpotQA
  • hotpot_50_qa_pairs.json - Corresponding question-answer pairs

Benchmarks

Each benchmark can be run independently with appropriate dependencies:

Mem0

pip install mem0ai openai
python qa_benchmark_mem0.py

LightRAG

pip install "lightrag-hku[api]"
python qa_benchmark_lightrag.py

Graphiti

pip install graphiti-core
python qa_benchmark_graphiti.py

Environment

Create .env with required API keys:

  • OPENAI_API_KEY (all benchmarks)
  • NEO4J_URI, NEO4J_USER, NEO4J_PASSWORD (Graphiti only)

Usage

Each benchmark inherits from QABenchmarkRAG base class and can be configured independently.

Results

Updated results will be posted soon.