From 67585d0ab1eee7151547d092671469d06d1041b2 Mon Sep 17 00:00:00 2001 From: Igor Ilic Date: Fri, 13 Dec 2024 15:30:24 +0100 Subject: [PATCH] feat: Add simple instruction for system prompt Add simple instruction for system prompt Feature COG-656 --- cognee/infrastructure/llm/prompts/answer_simple_question.txt | 1 + cognee/tasks/compute/query_compute.py | 2 +- 2 files changed, 2 insertions(+), 1 deletion(-) create mode 100644 cognee/infrastructure/llm/prompts/answer_simple_question.txt diff --git a/cognee/infrastructure/llm/prompts/answer_simple_question.txt b/cognee/infrastructure/llm/prompts/answer_simple_question.txt new file mode 100644 index 000000000..351e1e5e9 --- /dev/null +++ b/cognee/infrastructure/llm/prompts/answer_simple_question.txt @@ -0,0 +1 @@ +Answer the question using the provided context. Be as brief as possible. \ No newline at end of file diff --git a/cognee/tasks/compute/query_compute.py b/cognee/tasks/compute/query_compute.py index 1686701fe..cd606ddc5 100644 --- a/cognee/tasks/compute/query_compute.py +++ b/cognee/tasks/compute/query_compute.py @@ -24,7 +24,7 @@ async def query_compute(query: str) -> list: "context": found_chunks[0].payload["text"], } user_prompt = render_prompt("context_for_question.txt", args) - system_prompt = read_query_prompt("answer_hotpot_using_cognee_search.txt") + system_prompt = read_query_prompt("answer_simple_question.txt") llm_client = get_llm_client() computed_answer = await llm_client.acreate_structured_output(