LightRAG/k8s-deploy/lightrag-minimal/values-dev.yaml
Taddeus a70ba1f75a
Phase 1: LightRAG Minimal Helm chart and documentation indexing using url references (#2)
* Partial implementation of phase-0

* Partial implementation of phase-1

* add report

* add postgress

* Revert "add postgress"

This reverts commit 27778dc6bb3906b5220dd386e47fe32ca7415332.

* remove junk

* Cleaned up annd setup docs

* update docs

* moved report

* Updated load_markdown_files function: Now returns tuples with (content, title, relative_path) instead of just (content, title)

* fixes to load docs script and more env variables for llm configuration

* update prod values

* update docs

* apolo docs support with linking

* update docs to reflect url conventions and mapping with docs

* Adds ingress and forwardAuth configurations

Adds ingress configuration to expose the application.

Adds forwardAuth configuration to enable user authentication.

Includes middleware to strip headers.

* Adds ingress and forward authentication middleware support
2025-06-23 20:04:34 +03:00

75 lines
No EOL
1.6 KiB
YAML

# Development/Minikube Values
# Optimized for local development with reduced resource requirements
# Environment configuration
env:
LLM_MODEL: "gpt-4o"
WEBUI_TITLE: "Apolo Copilot - LightRAG (Development)"
WEBUI_DESCRIPTION: "Development LightRAG for Apolo Documentation"
# OpenAI API Configuration
LLM_BINDING: "openai"
LLM_BINDING_HOST: "https://api.openai.com/v1"
EMBEDDING_BINDING: "openai"
EMBEDDING_BINDING_HOST: "https://api.openai.com/v1"
EMBEDDING_MODEL: "text-embedding-ada-002"
EMBEDDING_DIM: "1536"
# Concurrency settings (conservative for API stability)
MAX_ASYNC: "4"
MAX_PARALLEL_INSERT: "2"
# LLM Configuration
ENABLE_LLM_CACHE: "true"
ENABLE_LLM_CACHE_FOR_EXTRACT: "true"
TIMEOUT: "240"
TEMPERATURE: "0"
MAX_TOKENS: "32768"
# Reduced resources for local development
resources:
limits:
cpu: 1000m
memory: 2Gi
requests:
cpu: 250m
memory: 512Mi
# Smaller storage for development
persistence:
ragStorage:
size: 5Gi
inputs:
size: 2Gi
# PostgreSQL with reduced resources
postgresql:
# Use pgvector image for vector support
image:
registry: docker.io
repository: pgvector/pgvector
tag: pg16
auth:
password: "dev-lightrag-pass"
primary:
persistence:
size: 5Gi
resources:
limits:
cpu: 500m
memory: 1Gi
requests:
cpu: 100m
memory: 256Mi
# OpenAI API key (set via environment variable)
secrets:
openaiApiKey: "${OPENAI_API_KEY}"
# Disable ingress for local development (use port-forward)
ingress:
enabled: false
# Disable autoscaling for development
autoscaling:
enabled: false