Commit graph

101 commits

Author SHA1 Message Date
Raphaël MANSUY
cf0899c063 cherry-pick 9d69e8d7 2025-12-04 19:19:00 +08:00
Raphaël MANSUY
3f309105b0 cherry-pick d5bcd14c 2025-12-04 19:18:40 +08:00
Raphaël MANSUY
ec0d9bd763 cherry-pick ee45ab51 2025-12-04 19:18:35 +08:00
Raphaël MANSUY
5bdd741eed cherry-pick 8bf41131 2025-12-04 19:18:34 +08:00
Raphaël MANSUY
4316172fbf cherry-pick 12facac5 2025-12-04 19:18:15 +08:00
Raphaël MANSUY
593b277945 cherry-pick 9f44e89d 2025-12-04 19:18:14 +08:00
Raphaël MANSUY
fae72fab75 cherry-pick 754d2ad2 2025-12-04 19:14:31 +08:00
anouarbm
7ce251c319 docs: Add documentation and examples for include_chunk_content parameter
Added comprehensive documentation for the new include_chunk_content parameter
that enables retrieval of actual chunk text content in API responses.

Documentation Updates:
- Added "Include Chunk Content in References" section to API README
- Explained use cases: RAG evaluation, debugging, citations, transparency
- Provided JSON request/response examples
- Clarified parameter interaction with include_references

OpenAPI/Swagger Examples:
- Added "Response with chunk content" example to /query endpoint
- Shows complete reference structure with content field
- Demonstrates realistic chunk text content

This makes the feature discoverable through:
1. API documentation (README.md)
2. Interactive Swagger UI (http://localhost:9621/docs)
3. Code examples for developers

(cherry picked from commit 963ad4c637)
2025-12-04 19:11:20 +08:00
xiaojunxiang
355aa2593c fix(docs): correct typo "acivate" → "activate"
(cherry picked from commit 9e5004e24f)
2025-12-04 19:11:08 +08:00
yangdx
db508954d1 Add uv package manager support to installation docs
(cherry picked from commit 7bc6ccea19)
2025-12-04 19:09:04 +08:00
Anush008
e86aa091f4 refactor: Qdrant Multi-tenancy (Include staged)
Signed-off-by: Anush008 <anushshetty90@gmail.com>
(cherry picked from commit 8584980e3a)
2025-12-04 19:09:01 +08:00
yangdx
b0bdbb5839 Add offline deployment support with cache management and layered deps
• Add tiktoken cache downloader CLI
• Add layered offline dependencies
• Add offline requirements files
• Add offline deployment guide

(cherry picked from commit a5c05f1b92)
2025-12-04 19:07:09 +08:00
yangdx
fcbab77613 docs: add API key usage example and security guidance 2025-09-28 10:56:52 +08:00
yangdx
6b953fa53d Remove auto-scan-at-startup feature and related documentation
• Remove --auto-scan-at-startup arg
• Delete auto scan docs sections
• Remove startup scanning logic
2025-09-23 16:24:53 +08:00
yangdx
7b371309dd Update README 2025-09-15 12:31:39 +08:00
yangdx
49ea9a79a7 Update rerank doc in README 2025-08-23 23:06:10 +08:00
yangdx
8c6b5f4a3a Update README 2025-08-21 18:14:27 +08:00
yangdx
62cdc7d7eb Update documentation with LLM selection guidelines and API improvements 2025-08-21 13:59:14 +08:00
yangdx
4c556d8aae Set default TIMEOUT value to 150, and gunicorn timeout to TIMEOUT+30 2025-08-20 22:04:32 +08:00
yangdx
3a7310873c Merge branch 'bedrock-support' 2025-08-17 02:23:44 +08:00
yangdx
da7e4b79e5 Update documentation in README files 2025-08-17 02:23:14 +08:00
yangdx
dc7a6e1c5b Update README 2025-08-16 06:15:27 +08:00
yangdx
0b5c708660 Update storage implementation documentation
- Add detailed storage type descriptions
- Remove Chroma from vector storage options
- Include recommended PostgreSQL version
- Add Memgraph to graph storage options
- Update performance comparison notes
2025-08-05 18:03:51 +08:00
yangdx
adf7ec8e35 feat: Add OpenAI LLM Options support with BindingOptions framework
- Add OpenAILLMOptions dataclass with full OpenAI API parameter support
- Integrate OpenAI options in config.py for automatic binding detection
- Update server functions to inject OpenAI options for openai/azure_openai bindings
2025-08-05 03:47:26 +08:00
yangdx
9d5603d35e Set the default LLM temperature to 1.0 and centralize constant management 2025-07-31 17:15:10 +08:00
yangdx
b456bb0942 Fix linting 2025-07-30 13:59:50 +08:00
yangdx
a788899285 Fix linting 2025-07-30 13:32:15 +08:00
yangdx
797dcc1ff1 Update README 2025-07-30 13:31:47 +08:00
yangdx
50621d5a94 Update docs 2025-07-30 13:16:26 +08:00
yangdx
75d1b1e9f8 Update Ollama context length configuration
- Rename OLLAMA_NUM_CTX to OLLAMA_LLM_NUM_CTX
- Increase default context window size
- Add requirement for minimum context size
- Update documentation examples
2025-07-29 09:53:37 +08:00
Michele Comitini
bd94714b15 options needs to be passed to ollama client embed() method
Fix line length

Create binding_options.py

Remove test property

Add dynamic binding options to CLI and environment config

Automatically generate command-line arguments and environment variable
support for all LLM provider bindings using BindingOptions. Add sample
.env generation and extensible framework for new providers.

Add example option definitions and fix test arg check in OllamaOptions

Add options_dict method to BindingOptions for argument parsing

Add comprehensive Ollama binding configuration options

ruff formatting Apply ruff formatting to binding_options.py

Add Ollama separate options for embedding and LLM

Refactor Ollama binding options and fix class var handling

The changes improve how class variables are handled in binding options
and better organize the Ollama-specific options into LLM and embedding
subclasses.

Fix typo in arg test.

Rename cls parameter to klass to avoid keyword shadowing

Fix Ollama embedding binding name typo

Fix ollama embedder context param name

Split Ollama options into LLM and embedding configs with mixin base

Add Ollama option configuration to LLM and embeddings in lightrag_server

Update sample .env generation and environment handling

Conditionally add env vars and cmdline options only when ollama bindings
are used. Add example env file for Ollama binding options.
2025-07-28 12:05:40 +02:00
yangdx
3f5ade47cd Update README 2025-07-27 17:26:49 +08:00
yangdx
0dfbce0bb4 Update the README to clarify the explanation of concurrent processes. 2025-07-27 10:39:28 +08:00
yangdx
b3c2987006 Reduce default MAX_TOKENS from 32000 to 10000 2025-07-26 08:13:49 +08:00
yangdx
51231c7647 Update README 2025-07-24 15:48:49 +08:00
yangdx
e828539b24 Update README 2025-07-17 19:05:34 +08:00
Daniel.y
6d1260aafa
Merge pull request #1766 from HKUDS/fix-memgraph-max-nodes-issue
Fix Memgraph get_knowledge_graph issues
2025-07-15 16:07:04 +08:00
yangdx
b03bb48e24 feat: Refine summary logic and add dedicated Ollama num_ctx config
- Refactor the trigger condition for LLM-based summarization of entities and relations. Instead of relying on character length, the summary is now triggered when the number of merged description fragments exceeds a configured threshold. This provides a more robust and logical condition for consolidation.
- Introduce the `OLLAMA_NUM_CTX` environment variable to explicitly configure the context window size (`num_ctx`) for Ollama models. This decouples the model's context length from the `MAX_TOKENS` parameter, which is now specifically used to limit input for summary generation, making the configuration clearer and more flexible.
- Updated `README` files, `env.example`, and default values to reflect these changes.
2025-07-14 01:55:04 +08:00
DavIvek
a0c4d88b0d wip fix Memgraph get_knowledge_graph issues 2025-07-10 16:56:44 +02:00
yangdx
f86ae6df0a Update api server README 2025-07-07 17:16:14 +08:00
yangdx
070a5db801 Update README 2025-05-23 12:50:48 +08:00
yangdx
c41f8d9ed3 Update README 2025-05-16 09:05:50 +08:00
yangdx
e5b0807298 Update README 2025-05-15 17:36:45 +08:00
yangdx
b9c25dfeb0 Update README 2025-05-14 14:42:52 +08:00
yangdx
db125c3764 Update README 2025-05-14 11:29:46 +08:00
yangdx
8145b436c8 Fix linting 2025-05-09 11:52:10 +08:00
yangdx
0751382e65 Update README.md 2025-05-09 11:51:22 +08:00
yangdx
31def4280e Update README 2025-04-16 15:36:44 +08:00
yangdx
1e79257976 Update README 2025-04-11 17:03:46 +08:00
earayu
224700466b fix grammar and typo in lightrag-api REAEME.md 2025-04-10 12:22:20 +08:00