diff --git a/env.example b/env.example index d86b98f7..4c8d355d 100644 --- a/env.example +++ b/env.example @@ -29,7 +29,7 @@ WEBUI_DESCRIPTION="Simple and Fast Graph Based RAG System" # OLLAMA_EMULATING_MODEL_NAME=lightrag OLLAMA_EMULATING_MODEL_TAG=latest -### Max nodes return from grap retrieval in webui +### Max nodes return from graph retrieval in webui # MAX_GRAPH_NODES=1000 ### Logging level @@ -56,22 +56,22 @@ OLLAMA_EMULATING_MODEL_TAG=latest ###################################################################################### ### Query Configuration ### -### How to control the context lenght sent to LLM: +### How to control the context length sent to LLM: ### MAX_ENTITY_TOKENS + MAX_RELATION_TOKENS < MAX_TOTAL_TOKENS -### Chunk_Tokens = MAX_TOTAL_TOKENS - Actual_Entity_Tokens - Actual_Reation_Tokens +### Chunk_Tokens = MAX_TOTAL_TOKENS - Actual_Entity_Tokens - Actual_Relation_Tokens ###################################################################################### -# LLM responde cache for query (Not valid for streaming response) +# LLM response cache for query (Not valid for streaming response) ENABLE_LLM_CACHE=true # COSINE_THRESHOLD=0.2 ### Number of entities or relations retrieved from KG # TOP_K=40 -### Maxmium number or chunks for naive vector search +### Maximum number or chunks for naive vector search # CHUNK_TOP_K=20 -### control the actual enties send to LLM +### control the actual entities send to LLM # MAX_ENTITY_TOKENS=6000 ### control the actual relations send to LLM # MAX_RELATION_TOKENS=8000 -### control the maximum tokens send to LLM (include entities, raltions and chunks) +### control the maximum tokens send to LLM (include entities, relations and chunks) # MAX_TOTAL_TOKENS=30000 ### maximum number of related chunks per source entity or relation @@ -93,7 +93,7 @@ ENABLE_LLM_CACHE=true RERANK_BINDING=null ### Enable rerank by default in query params when RERANK_BINDING is not null # RERANK_BY_DEFAULT=True -### rerank score chunk filter(set to 0.0 to keep all chunks, 0.6 or above if LLM is not strong enought) +### rerank score chunk filter(set to 0.0 to keep all chunks, 0.6 or above if LLM is not strong enough) # MIN_RERANK_SCORE=0.0 ### For local deployment with vLLM @@ -131,7 +131,7 @@ SUMMARY_LANGUAGE=English # CHUNK_SIZE=1200 # CHUNK_OVERLAP_SIZE=100 -### Number of summary semgments or tokens to trigger LLM summary on entity/relation merge (at least 3 is recommented) +### Number of summary segments or tokens to trigger LLM summary on entity/relation merge (at least 3 is recommended) # FORCE_LLM_SUMMARY_ON_MERGE=8 ### Max description token size to trigger LLM summary # SUMMARY_MAX_TOKENS = 1200 @@ -179,7 +179,7 @@ LLM_BINDING_API_KEY=your_api_key # OPENAI_LLM_TEMPERATURE=0.9 ### Set the max_tokens to mitigate endless output of some LLM (less than LLM_TIMEOUT * llm_output_tokens/second, i.e. 9000 = 180s * 50 tokens/s) ### Typically, max_tokens does not include prompt content, though some models, such as Gemini Models, are exceptions -### For vLLM/SGLang doployed models, or most of OpenAI compatible API provider +### For vLLM/SGLang deployed models, or most of OpenAI compatible API provider # OPENAI_LLM_MAX_TOKENS=9000 ### For OpenAI o1-mini or newer modles OPENAI_LLM_MAX_COMPLETION_TOKENS=9000 @@ -193,7 +193,7 @@ OPENAI_LLM_MAX_COMPLETION_TOKENS=9000 # OPENAI_LLM_REASONING_EFFORT=minimal ### OpenRouter Specific Parameters # OPENAI_LLM_EXTRA_BODY='{"reasoning": {"enabled": false}}' -### Qwen3 Specific Parameters depoly by vLLM +### Qwen3 Specific Parameters deploy by vLLM # OPENAI_LLM_EXTRA_BODY='{"chat_template_kwargs": {"enable_thinking": false}}' ### use the following command to see all support options for Ollama LLM @@ -247,8 +247,8 @@ OLLAMA_EMBEDDING_NUM_CTX=8192 ### lightrag-server --embedding-binding ollama --help #################################################################### -### WORKSPACE setting workspace name for all storage types -### in the purpose of isolating data from LightRAG instances. +### WORKSPACE sets workspace name for all storage types +### for the purpose of isolating data from LightRAG instances. ### Valid workspace name constraints: a-z, A-Z, 0-9, and _ #################################################################### # WORKSPACE=space1