Added docs
This commit is contained in:
parent
ec3b453ce3
commit
e925867a7d
1 changed files with 5 additions and 11 deletions
|
|
@ -28,22 +28,19 @@ With the integration of Keepi.ai, we encountered several challenges that made us
|
||||||
|
|
||||||
Moving forward, we have adopted several new strategies, features, and design principles:
|
Moving forward, we have adopted several new strategies, features, and design principles:
|
||||||
|
|
||||||
<aside>
|
|
||||||
💡 Propositions: Defined as atomic expressions within a text, each proposition encapsulates a unique factoid, conveyed in a succinct, standalone natural language format.
|
|
||||||
|
|
||||||
|
### Propositions:
|
||||||
|
|
||||||
|
Defined as atomic expressions within a text, each proposition encapsulates a unique factoid, conveyed in a succinct, standalone natural language format.
|
||||||
We employ Large Language Models (LLMs) to break down text into propositions and link them, forming graphs with propositions as nodes and their connections as edges.
|
We employ Large Language Models (LLMs) to break down text into propositions and link them, forming graphs with propositions as nodes and their connections as edges.
|
||||||
For example, "Grass is green", and "2 + 5 = 5" are propositions. The first proposition has the truth value of "true" and the second "false".
|
For example, "Grass is green", and "2 + 5 = 5" are propositions. The first proposition has the truth value of "true" and the second "false".
|
||||||
The inspiration was found in the following paper: https://arxiv.org/pdf/2312.06648.pdf
|
The inspiration was found in the following paper: https://arxiv.org/pdf/2312.06648.pdf
|
||||||
|
|
||||||
</aside>
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
<aside>
|
|
||||||
|
|
||||||
|
### Multilayer Graph Network:
|
||||||
|
|
||||||
💡 Multilayer Graph Network:
|
|
||||||
|
|
||||||
A cognitive multilayer networks is both a quantitative and interpretive framework for exploring the mental lexicon, the intricate cognitive system that stores information about known words/concepts.
|
A cognitive multilayer networks is both a quantitative and interpretive framework for exploring the mental lexicon, the intricate cognitive system that stores information about known words/concepts.
|
||||||
|
|
||||||
|
|
@ -55,18 +52,15 @@ For example, if "John Doe" authored two New York Times cooking articles, we coul
|
||||||
|
|
||||||
We used concepts from psycholinguistics described here: https://arxiv.org/abs/1507.08539
|
We used concepts from psycholinguistics described here: https://arxiv.org/abs/1507.08539
|
||||||
|
|
||||||
</aside>
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
<aside>
|
### Data Loader:
|
||||||
💡 Data Loader:
|
|
||||||
|
|
||||||
It’s vital that we address the technical challenges associated with Retrieval-Augmented Generation (RAG), such as metadata management, context retrieval, knowledge sanitization, and data enrichment.
|
It’s vital that we address the technical challenges associated with Retrieval-Augmented Generation (RAG), such as metadata management, context retrieval, knowledge sanitization, and data enrichment.
|
||||||
|
|
||||||
The solution lies in a dependable data pipeline capable of efficiently and scalably preparing and loading data in various formats from a range of different sources. For this purpose, we can use 'dlt' as our data loader, gaining access to over 28 supported data sources.
|
The solution lies in a dependable data pipeline capable of efficiently and scalably preparing and loading data in various formats from a range of different sources. For this purpose, we can use 'dlt' as our data loader, gaining access to over 28 supported data sources.
|
||||||
|
|
||||||
</aside>
|
|
||||||
|
|
||||||
To enhance the Pythonic interface, we streamlined the use of cognee into three primary methods. Users can now execute the following steps:
|
To enhance the Pythonic interface, we streamlined the use of cognee into three primary methods. Users can now execute the following steps:
|
||||||
|
|
||||||
|
|
|
||||||
Loading…
Add table
Reference in a new issue