<!-- .github/pull_request_template.md --> This PR contains the evaluation framework development for cognee ## DCO Affirmation I affirm that all code in every commit of this pull request conforms to the terms of the Topoteretes Developer Certificate of Origin <!-- This is an auto-generated comment: release notes by coderabbit.ai --> ## Summary by CodeRabbit - **New Features** - Expanded evaluation framework now integrates asynchronous corpus building, question answering, and performance evaluation with adaptive benchmarks for improved metrics (correctness, exact match, and F1 score). - **Infrastructure** - Added database integration for persistent storage of questions, answers, and metrics. - Launched an interactive metrics dashboard featuring advanced visualizations. - Introduced an automated testing workflow for continuous quality assurance. - **Documentation** - Updated guidelines for generating concise, clear answers. <!-- end of auto-generated comment: release notes by coderabbit.ai -->
15 lines
439 B
Python
15 lines
439 B
Python
from datetime import datetime, timezone
|
|
from uuid import uuid4
|
|
from sqlalchemy import Column, DateTime, JSON, UUID
|
|
|
|
from cognee.modules.data.models.answers_base import AnswersBase
|
|
|
|
|
|
class Answers(AnswersBase):
|
|
__tablename__ = "eval_answers"
|
|
|
|
id = Column(UUID, primary_key=True, default=uuid4)
|
|
|
|
payload = Column(JSON, nullable=False)
|
|
|
|
created_at = Column(DateTime(timezone=True), default=lambda: datetime.now(timezone.utc))
|