Update README
This commit is contained in:
parent
9b63a89054
commit
3c530b21b6
3 changed files with 9 additions and 5 deletions
|
|
@ -320,7 +320,7 @@ class QueryParam:
|
||||||
max_relation_tokens: int = int(os.getenv("MAX_RELATION_TOKENS", "10000"))
|
max_relation_tokens: int = int(os.getenv("MAX_RELATION_TOKENS", "10000"))
|
||||||
"""Maximum number of tokens allocated for relationship context in unified token control system."""
|
"""Maximum number of tokens allocated for relationship context in unified token control system."""
|
||||||
|
|
||||||
max_total_tokens: int = int(os.getenv("MAX_TOTAL_TOKENS", "32000"))
|
max_total_tokens: int = int(os.getenv("MAX_TOTAL_TOKENS", "30000"))
|
||||||
"""Maximum total tokens budget for the entire query context (entities + relations + chunks + system prompt)."""
|
"""Maximum total tokens budget for the entire query context (entities + relations + chunks + system prompt)."""
|
||||||
|
|
||||||
hl_keywords: list[str] = field(default_factory=list)
|
hl_keywords: list[str] = field(default_factory=list)
|
||||||
|
|
|
||||||
|
|
@ -327,7 +327,7 @@ class QueryParam:
|
||||||
max_relation_tokens: int = int(os.getenv("MAX_RELATION_TOKENS", "10000"))
|
max_relation_tokens: int = int(os.getenv("MAX_RELATION_TOKENS", "10000"))
|
||||||
"""Maximum number of tokens allocated for relationship context in unified token control system."""
|
"""Maximum number of tokens allocated for relationship context in unified token control system."""
|
||||||
|
|
||||||
max_total_tokens: int = int(os.getenv("MAX_TOTAL_TOKENS", "32000"))
|
max_total_tokens: int = int(os.getenv("MAX_TOTAL_TOKENS", "30000"))
|
||||||
"""Maximum total tokens budget for the entire query context (entities + relations + chunks + system prompt)."""
|
"""Maximum total tokens budget for the entire query context (entities + relations + chunks + system prompt)."""
|
||||||
|
|
||||||
conversation_history: list[dict[str, str]] = field(default_factory=list)
|
conversation_history: list[dict[str, str]] = field(default_factory=list)
|
||||||
|
|
|
||||||
10
env.example
10
env.example
|
|
@ -48,15 +48,19 @@ OLLAMA_EMULATING_MODEL_TAG=latest
|
||||||
# LIGHTRAG_API_KEY=your-secure-api-key-here
|
# LIGHTRAG_API_KEY=your-secure-api-key-here
|
||||||
# WHITELIST_PATHS=/health,/api/*
|
# WHITELIST_PATHS=/health,/api/*
|
||||||
|
|
||||||
########################
|
######################################################################################
|
||||||
### Query Configuration
|
### Query Configuration
|
||||||
########################
|
###
|
||||||
|
### How to control the context lenght sent to LLM:
|
||||||
|
### MAX_ENTITY_TOKENS + MAX_RELATION_TOKENS < MAX_TOTAL_TOKENS
|
||||||
|
### Chunk_Tokens = MAX_TOTAL_TOKENS - Actual_Entity_Tokens - Actual_Reation_Tokens
|
||||||
|
######################################################################################
|
||||||
# LLM responde cache for query (Not valid for streaming response)
|
# LLM responde cache for query (Not valid for streaming response)
|
||||||
ENABLE_LLM_CACHE=true
|
ENABLE_LLM_CACHE=true
|
||||||
# COSINE_THRESHOLD=0.2
|
# COSINE_THRESHOLD=0.2
|
||||||
### Number of entities or relations retrieved from KG
|
### Number of entities or relations retrieved from KG
|
||||||
# TOP_K=40
|
# TOP_K=40
|
||||||
### Maxmium number or chunks plan to send to LLM
|
### Maxmium number or chunks for naive vactor search
|
||||||
# CHUNK_TOP_K=10
|
# CHUNK_TOP_K=10
|
||||||
### control the actual enties send to LLM
|
### control the actual enties send to LLM
|
||||||
# MAX_ENTITY_TOKENS=10000
|
# MAX_ENTITY_TOKENS=10000
|
||||||
|
|
|
||||||
Loading…
Add table
Reference in a new issue