diff --git a/.github/workflows/build-multiarch.yml b/.github/workflows/build-multiarch.yml index b8ee1a28..f9a83400 100644 --- a/.github/workflows/build-multiarch.yml +++ b/.github/workflows/build-multiarch.yml @@ -53,58 +53,63 @@ jobs: # backend - image: backend file: ./Dockerfile.backend - tag: phact/openrag-backend + tag: langflowai/openrag-backend platform: linux/amd64 arch: amd64 runs-on: ubuntu-latest-16-cores - image: backend file: ./Dockerfile.backend - tag: phact/openrag-backend + tag: langflowai/openrag-backend platform: linux/arm64 arch: arm64 - runs-on: [self-hosted, linux, ARM64, langflow-ai-arm64-2] + #runs-on: [self-hosted, linux, ARM64, langflow-ai-arm64-2] + runs-on: RagRunner # frontend - image: frontend file: ./Dockerfile.frontend - tag: phact/openrag-frontend + tag: langflowai/openrag-frontend platform: linux/amd64 arch: amd64 runs-on: ubuntu-latest-16-cores - image: frontend file: ./Dockerfile.frontend - tag: phact/openrag-frontend + tag: langflowai/openrag-frontend platform: linux/arm64 arch: arm64 - runs-on: [self-hosted, linux, ARM64, langflow-ai-arm64-2] + #runs-on: [self-hosted, linux, ARM64, langflow-ai-arm64-2] + runs-on: RagRunner # langflow - image: langflow file: ./Dockerfile.langflow - tag: phact/openrag-langflow + tag: langflowai/openrag-langflow platform: linux/amd64 arch: amd64 runs-on: ubuntu-latest-16-cores - image: langflow file: ./Dockerfile.langflow - tag: phact/openrag-langflow + tag: langflowai/openrag-langflow platform: linux/arm64 arch: arm64 - runs-on: self-hosted + #runs-on: self-hosted + runs-on: RagRunner # opensearch - image: opensearch file: ./Dockerfile - tag: phact/openrag-opensearch + tag: langflowai/openrag-opensearch platform: linux/amd64 arch: amd64 runs-on: ubuntu-latest-16-cores - image: opensearch file: ./Dockerfile - tag: phact/openrag-opensearch + tag: langflowai/openrag-opensearch platform: linux/arm64 arch: arm64 - runs-on: [self-hosted, linux, ARM64, langflow-ai-arm64-2] + #runs-on: [self-hosted, linux, ARM64, langflow-ai-arm64-2] + #runs-on: self-hosted + runs-on: RagRunner runs-on: ${{ matrix.runs-on }} @@ -165,40 +170,40 @@ jobs: VERSION=${{ steps.version.outputs.version }} # Create versioned tags - docker buildx imagetools create -t phact/openrag-backend:$VERSION \ - phact/openrag-backend:$VERSION-amd64 \ - phact/openrag-backend:$VERSION-arm64 + docker buildx imagetools create -t langflowai/openrag-backend:$VERSION \ + langflowai/openrag-backend:$VERSION-amd64 \ + langflowai/openrag-backend:$VERSION-arm64 - docker buildx imagetools create -t phact/openrag-frontend:$VERSION \ - phact/openrag-frontend:$VERSION-amd64 \ - phact/openrag-frontend:$VERSION-arm64 + docker buildx imagetools create -t langflowai/openrag-frontend:$VERSION \ + langflowai/openrag-frontend:$VERSION-amd64 \ + langflowai/openrag-frontend:$VERSION-arm64 - docker buildx imagetools create -t phact/openrag-langflow:$VERSION \ - phact/openrag-langflow:$VERSION-amd64 \ - phact/openrag-langflow:$VERSION-arm64 + docker buildx imagetools create -t langflowai/openrag-langflow:$VERSION \ + langflowai/openrag-langflow:$VERSION-amd64 \ + langflowai/openrag-langflow:$VERSION-arm64 - docker buildx imagetools create -t phact/openrag-opensearch:$VERSION \ - phact/openrag-opensearch:$VERSION-amd64 \ - phact/openrag-opensearch:$VERSION-arm64 + docker buildx imagetools create -t langflowai/openrag-opensearch:$VERSION \ + langflowai/openrag-opensearch:$VERSION-amd64 \ + langflowai/openrag-opensearch:$VERSION-arm64 # Only update latest tags if version is numeric if [[ "$VERSION" =~ ^[0-9.-]+$ ]]; then echo "Updating latest tags for production release: $VERSION" - docker buildx imagetools create -t phact/openrag-backend:latest \ - phact/openrag-backend:$VERSION-amd64 \ - phact/openrag-backend:$VERSION-arm64 + docker buildx imagetools create -t langflowai/openrag-backend:latest \ + langflowai/openrag-backend:$VERSION-amd64 \ + langflowai/openrag-backend:$VERSION-arm64 - docker buildx imagetools create -t phact/openrag-frontend:latest \ - phact/openrag-frontend:$VERSION-amd64 \ - phact/openrag-frontend:$VERSION-arm64 + docker buildx imagetools create -t langflowai/openrag-frontend:latest \ + langflowai/openrag-frontend:$VERSION-amd64 \ + langflowai/openrag-frontend:$VERSION-arm64 - docker buildx imagetools create -t phact/openrag-langflow:latest \ - phact/openrag-langflow:$VERSION-amd64 \ - phact/openrag-langflow:$VERSION-arm64 + docker buildx imagetools create -t langflowai/openrag-langflow:latest \ + langflowai/openrag-langflow:$VERSION-amd64 \ + langflowai/openrag-langflow:$VERSION-arm64 - docker buildx imagetools create -t phact/openrag-opensearch:latest \ - phact/openrag-opensearch:$VERSION-amd64 \ - phact/openrag-opensearch:$VERSION-arm64 + docker buildx imagetools create -t langflowai/openrag-opensearch:latest \ + langflowai/openrag-opensearch:$VERSION-amd64 \ + langflowai/openrag-opensearch:$VERSION-arm64 else echo "Skipping latest tags - version: $VERSION (not numeric)" fi diff --git a/.github/workflows/test-integration.yml b/.github/workflows/test-integration.yml index 1b4c6162..a70dd24d 100644 --- a/.github/workflows/test-integration.yml +++ b/.github/workflows/test-integration.yml @@ -31,14 +31,23 @@ jobs: steps: - run: df -h - #- name: "node-cleanup" - #run: | - # sudo rm -rf /usr/share/dotnet /usr/local/lib/android /opt/ghc /opt/hostedtoolcache/CodeQL - # sudo docker image prune --all --force - # sudo docker builder prune -a + + - name: Cleanup Docker cache + run: | + docker system prune -af || true + docker builder prune -af || true + docker-compose -f docker-compose.yml down -v --remove-orphans || true + - run: df -h + - name: Checkout uses: actions/checkout@v4 + + - name: Verify workspace + run: | + echo "Current directory: $(pwd)" + echo "Workspace: ${GITHUB_WORKSPACE}" + ls -la - name: Set up UV uses: astral-sh/setup-uv@v3 diff --git a/Dockerfile.backend b/Dockerfile.backend index 5d9d84f4..3c94a279 100644 --- a/Dockerfile.backend +++ b/Dockerfile.backend @@ -21,7 +21,7 @@ COPY pyproject.toml uv.lock ./ RUN uv sync # Copy sample document and warmup script for docling -COPY documents/warmup_ocr.pdf ./ +COPY openrag-documents/warmup_ocr.pdf ./ COPY warm_up_docling.py ./ RUN uv run docling-tools models download RUN uv run python - <<'PY' diff --git a/Dockerfile.langflow b/Dockerfile.langflow index 5343585a..aaa9f5d6 100644 --- a/Dockerfile.langflow +++ b/Dockerfile.langflow @@ -1,4 +1,4 @@ -FROM langflowai/langflow-nightly:1.7.0.dev5 +FROM langflowai/langflow-nightly:1.7.0.dev21 EXPOSE 7860 diff --git a/Makefile b/Makefile index f47bba28..b5804f77 100644 --- a/Makefile +++ b/Makefile @@ -210,7 +210,7 @@ test-ci: echo "Pulling latest images..."; \ docker compose -f docker-compose-cpu.yml pull; \ echo "Building OpenSearch image override..."; \ - docker build --no-cache -t phact/openrag-opensearch:latest -f Dockerfile .; \ + docker build --no-cache -t langflowai/openrag-opensearch:latest -f Dockerfile .; \ echo "Starting infra (OpenSearch + Dashboards + Langflow) with CPU containers"; \ docker compose -f docker-compose-cpu.yml up -d opensearch dashboards langflow; \ echo "Starting docling-serve..."; \ @@ -288,10 +288,10 @@ test-ci-local: echo "Cleaning up old containers and volumes..."; \ docker compose -f docker-compose-cpu.yml down -v 2>/dev/null || true; \ echo "Building all images locally..."; \ - docker build -t phact/openrag-opensearch:latest -f Dockerfile .; \ - docker build -t phact/openrag-backend:latest -f Dockerfile.backend .; \ - docker build -t phact/openrag-frontend:latest -f Dockerfile.frontend .; \ - docker build -t phact/openrag-langflow:latest -f Dockerfile.langflow .; \ + docker build -t langflowai/openrag-opensearch:latest -f Dockerfile .; \ + docker build -t langflowai/openrag-backend:latest -f Dockerfile.backend .; \ + docker build -t langflowai/openrag-frontend:latest -f Dockerfile.frontend .; \ + docker build -t langflowai/openrag-langflow:latest -f Dockerfile.langflow .; \ echo "Starting infra (OpenSearch + Dashboards + Langflow) with CPU containers"; \ docker compose -f docker-compose-cpu.yml up -d opensearch dashboards langflow; \ echo "Starting docling-serve..."; \ diff --git a/README.md b/README.md index d7d1b614..4e861740 100644 --- a/README.md +++ b/README.md @@ -18,27 +18,29 @@ OpenRAG is a comprehensive Retrieval-Augmented Generation platform that enables
Quickstart   |   - TUI Interface   |   - Docker Deployment   |   + Python package   |   + Docker or Podman   |   Development   |   Troubleshooting
## Quickstart -To quickly run OpenRAG without creating or modifying any project files, use `uvx`: +To run OpenRAG without creating or modifying any project files, use `uvx`: ```bash uvx openrag ``` -This runs OpenRAG without installing it to your project or globally. -To run a specific version of OpenRAG, add the version to the command, such as: `uvx --from openrag==0.1.25 openrag`. + +This command runs OpenRAG without installing it to your project or globally. + +To run a specific version of OpenRAG, run `uvx --from openrag==VERSION openrag`. ## Install Python package -To first set up a project and then install the OpenRAG Python package, do the following: +To add the OpenRAG Python package to a Python project, use `uv`: -1. Create a new project with a virtual environment using `uv init`. +1. Create a new project with a virtual environment using `uv init`: ```bash uv init YOUR_PROJECT_NAME @@ -48,33 +50,33 @@ To first set up a project and then install the OpenRAG Python package, do the fo The `(venv)` prompt doesn't change, but `uv` commands will automatically use the project's virtual environment. For more information on virtual environments, see the [uv documentation](https://docs.astral.sh/uv/pip/environments). -2. Add OpenRAG to your project. +2. Add OpenRAG to your project: + ```bash uv add openrag ``` - To add a specific version of OpenRAG: - ```bash - uv add openrag==0.1.25 - ``` + To add a specific version of OpenRAG, run `uv add openrag==VERSION`. + +3. Start the OpenRAG terminal user interface (TUI): -3. Start the OpenRAG TUI. ```bash uv run openrag ``` 4. Continue with the [Quickstart](https://docs.openr.ag/quickstart). -For the full TUI installation guide, see [TUI](https://docs.openr.ag/install). +For all installation options, see the [OpenRAG installation guide](https://docs.openr.ag/install). ## Docker or Podman installation -For more information, see [Install OpenRAG containers](https://docs.openr.ag/docker). - -## Troubleshooting - -For common issues and fixes, see [Troubleshoot](https://docs.openr.ag/support/troubleshoot). +By default, OpenRAG automatically starts the required containers and helps you manage them. +To install OpenRAG with self-managed containers, see the [OpenRAG installation guide](https://docs.openr.ag/docker). ## Development -For developers wanting to contribute to OpenRAG or set up a development environment, see [CONTRIBUTING.md](CONTRIBUTING.md). \ No newline at end of file +For developers wanting to contribute to OpenRAG or set up a development environment, see [CONTRIBUTING.md](CONTRIBUTING.md). + +## Troubleshooting + +For common issues and fixes, see [Troubleshoot OpenRAG](https://docs.openr.ag/support/troubleshoot). \ No newline at end of file diff --git a/docker-compose-cpu.yml b/docker-compose-cpu.yml index 50e118b7..61b9a2f6 100644 --- a/docker-compose-cpu.yml +++ b/docker-compose-cpu.yml @@ -1,6 +1,6 @@ services: opensearch: - image: phact/openrag-opensearch:${OPENRAG_VERSION:-latest} + image: langflowai/openrag-opensearch:${OPENRAG_VERSION:-latest} #build: # context: . # dockerfile: Dockerfile @@ -44,7 +44,7 @@ services: - "5601:5601" openrag-backend: - image: phact/openrag-backend:${OPENRAG_VERSION:-latest} + image: langflowai/openrag-backend:${OPENRAG_VERSION:-latest} # build: # context: . # dockerfile: Dockerfile.backend @@ -81,12 +81,12 @@ services: - AWS_ACCESS_KEY_ID=${AWS_ACCESS_KEY_ID} - AWS_SECRET_ACCESS_KEY=${AWS_SECRET_ACCESS_KEY} volumes: - - ./documents:/app/documents:Z + - ./openrag-documents:/app/documents:Z - ./keys:/app/keys:Z - ./flows:/app/flows:U,z openrag-frontend: - image: phact/openrag-frontend:${OPENRAG_VERSION:-latest} + image: langflowai/openrag-frontend:${OPENRAG_VERSION:-latest} # build: # context: . # dockerfile: Dockerfile.frontend @@ -101,7 +101,7 @@ services: langflow: volumes: - ./flows:/app/flows:U,z - image: phact/openrag-langflow:${LANGFLOW_VERSION:-latest} + image: langflowai/openrag-langflow:${LANGFLOW_VERSION:-latest} # build: # context: . # dockerfile: Dockerfile.langflow @@ -129,7 +129,8 @@ services: - FILENAME=None - MIMETYPE=None - FILESIZE=0 - - LANGFLOW_VARIABLES_TO_GET_FROM_ENVIRONMENT=JWT,OPENRAG-QUERY-FILTER,OPENSEARCH_PASSWORD,OWNER,OWNER_NAME,OWNER_EMAIL,CONNECTOR_TYPE,FILENAME,MIMETYPE,FILESIZE + - SELECTED_EMBEDDING_MODEL=${SELECTED_EMBEDDING_MODEL:-} + - LANGFLOW_VARIABLES_TO_GET_FROM_ENVIRONMENT=JWT,OPENRAG-QUERY-FILTER,OPENSEARCH_PASSWORD,OWNER,OWNER_NAME,OWNER_EMAIL,CONNECTOR_TYPE,FILENAME,MIMETYPE,FILESIZE,SELECTED_EMBEDDING_MODEL - LANGFLOW_LOG_LEVEL=DEBUG - LANGFLOW_AUTO_LOGIN=${LANGFLOW_AUTO_LOGIN} - LANGFLOW_SUPERUSER=${LANGFLOW_SUPERUSER} diff --git a/docker-compose.gpu.yml b/docker-compose.gpu.yml new file mode 100644 index 00000000..4496c4ac --- /dev/null +++ b/docker-compose.gpu.yml @@ -0,0 +1,7 @@ +services: + openrag-backend: + environment: + - NVIDIA_DRIVER_CAPABILITIES=compute,utility + - NVIDIA_VISIBLE_DEVICES=all + gpus: all + diff --git a/docker-compose.yml b/docker-compose.yml index 7ba0cea8..2ed199a5 100644 --- a/docker-compose.yml +++ b/docker-compose.yml @@ -1,6 +1,6 @@ services: opensearch: - image: phact/openrag-opensearch:${OPENRAG_VERSION:-latest} + image: langflowai/openrag-opensearch:${OPENRAG_VERSION:-latest} #build: #context: . #dockerfile: Dockerfile @@ -44,10 +44,10 @@ services: - "5601:5601" openrag-backend: - image: phact/openrag-backend:${OPENRAG_VERSION:-latest} - # build: - # context: . - # dockerfile: Dockerfile.backend + image: langflowai/openrag-backend:${OPENRAG_VERSION:-latest} + build: + context: . + dockerfile: Dockerfile.backend container_name: openrag-backend depends_on: - langflow @@ -72,8 +72,6 @@ services: - WATSONX_ENDPOINT=${WATSONX_ENDPOINT} - WATSONX_PROJECT_ID=${WATSONX_PROJECT_ID} - OLLAMA_ENDPOINT=${OLLAMA_ENDPOINT} - - NVIDIA_DRIVER_CAPABILITIES=compute,utility - - NVIDIA_VISIBLE_DEVICES=all - GOOGLE_OAUTH_CLIENT_ID=${GOOGLE_OAUTH_CLIENT_ID} - GOOGLE_OAUTH_CLIENT_SECRET=${GOOGLE_OAUTH_CLIENT_SECRET} - MICROSOFT_GRAPH_OAUTH_CLIENT_ID=${MICROSOFT_GRAPH_OAUTH_CLIENT_ID} @@ -82,16 +80,15 @@ services: - AWS_ACCESS_KEY_ID=${AWS_ACCESS_KEY_ID} - AWS_SECRET_ACCESS_KEY=${AWS_SECRET_ACCESS_KEY} volumes: - - ./documents:/app/documents:Z + - ./openrag-documents:/app/documents:Z - ./keys:/app/keys:Z - ./flows:/app/flows:U,z - gpus: all openrag-frontend: - image: phact/openrag-frontend:${OPENRAG_VERSION:-latest} - # build: - # context: . - # dockerfile: Dockerfile.frontend + image: langflowai/openrag-frontend:${OPENRAG_VERSION:-latest} + build: + context: . + dockerfile: Dockerfile.frontend container_name: openrag-frontend depends_on: - openrag-backend @@ -103,21 +100,21 @@ services: langflow: volumes: - ./flows:/app/flows:U,z - image: phact/openrag-langflow:${LANGFLOW_VERSION:-latest} - # build: - # context: . - # dockerfile: Dockerfile.langflow + image: langflowai/openrag-langflow:${LANGFLOW_VERSION:-latest} + build: + context: . + dockerfile: Dockerfile.langflow container_name: langflow ports: - "7860:7860" environment: - LANGFLOW_DEACTIVATE_TRACING=true - - OPENAI_API_KEY=${OPENAI_API_KEY} - - ANTHROPIC_API_KEY=${ANTHROPIC_API_KEY} - - WATSONX_API_KEY=${WATSONX_API_KEY} - - WATSONX_ENDPOINT=${WATSONX_ENDPOINT} - - WATSONX_PROJECT_ID=${WATSONX_PROJECT_ID} - - OLLAMA_BASE_URL=${OLLAMA_ENDPOINT} + - OPENAI_API_KEY=${OPENAI_API_KEY:-None} + - ANTHROPIC_API_KEY=${ANTHROPIC_API_KEY:-None} + - WATSONX_API_KEY=${WATSONX_API_KEY:-None} + - WATSONX_ENDPOINT=${WATSONX_ENDPOINT:-None} + - WATSONX_PROJECT_ID=${WATSONX_PROJECT_ID:-None} + - OLLAMA_BASE_URL=${OLLAMA_ENDPOINT:-None} - LANGFLOW_LOAD_FLOWS_PATH=/app/flows - LANGFLOW_SECRET_KEY=${LANGFLOW_SECRET_KEY} - JWT=None @@ -127,11 +124,13 @@ services: - CONNECTOR_TYPE=system - CONNECTOR_TYPE_URL=url - OPENRAG-QUERY-FILTER="{}" + - OPENSEARCH_PASSWORD=${OPENSEARCH_PASSWORD} - FILENAME=None - MIMETYPE=None - FILESIZE=0 + - SELECTED_EMBEDDING_MODEL=${SELECTED_EMBEDDING_MODEL:-} - OPENSEARCH_PASSWORD=${OPENSEARCH_PASSWORD} - - LANGFLOW_VARIABLES_TO_GET_FROM_ENVIRONMENT=JWT,OPENRAG-QUERY-FILTER,OPENSEARCH_PASSWORD,OWNER,OWNER_NAME,OWNER_EMAIL,CONNECTOR_TYPE,FILENAME,MIMETYPE,FILESIZE + - LANGFLOW_VARIABLES_TO_GET_FROM_ENVIRONMENT=JWT,OPENRAG-QUERY-FILTER,OPENSEARCH_PASSWORD,OWNER,OWNER_NAME,OWNER_EMAIL,CONNECTOR_TYPE,FILENAME,MIMETYPE,FILESIZE,SELECTED_EMBEDDING_MODEL,OPENAI_API_KEY,ANTHROPIC_API_KEY,WATSONX_API_KEY,WATSONX_ENDPOINT,WATSONX_PROJECT_ID,OLLAMA_BASE_URL - LANGFLOW_LOG_LEVEL=DEBUG - LANGFLOW_AUTO_LOGIN=${LANGFLOW_AUTO_LOGIN} - LANGFLOW_SUPERUSER=${LANGFLOW_SUPERUSER} diff --git a/docs/README.md b/docs/README.md index 436d7830..bad5759c 100644 --- a/docs/README.md +++ b/docs/README.md @@ -42,7 +42,7 @@ If you are using GitHub pages for hosting, this command is a convenient way to b ## Update the OpenRAG documentation PDF -The documentation PDF at `openrag/documents/openrag-documentation.pdf` is used by the OpenRAG application, so keep it up to date. +The documentation PDF at `openrag/openrag-documents/openrag-documentation.pdf` is used by the OpenRAG application, so keep it up to date. To update the PDF, do the following: @@ -68,7 +68,7 @@ To remove these items, give the following prompt or something similar to your ID 2. Check your `.mdx` files to confirm these elements are removed. Don't commit the changes. -3. From `openrag/docs`, run this command to build the site with the changes, and create a PDF at `openrag/documents`. +3. From `openrag/docs`, run this command to build the site with the changes, and create a PDF at `openrag/openrag-documents`. ``` npm run build:pdf diff --git a/docs/docs/core-components/knowledge.mdx b/docs/docs/core-components/knowledge.mdx index ade572c6..0aac4313 100644 --- a/docs/docs/core-components/knowledge.mdx +++ b/docs/docs/core-components/knowledge.mdx @@ -140,7 +140,7 @@ The default value is 200 characters, which represents an overlap of 20 percent i ### Set the local documents path {#set-the-local-documents-path} -The default path for local uploads is the `./documents` subdirectory in your OpenRAG installation directory. This is mounted to the `/app/documents/` directory inside the OpenRAG container. Files added to the host or container directory are visible in both locations. +The default path for local uploads is the `./openrag-documents` subdirectory in your OpenRAG installation directory. This is mounted to the `/app/documents/` directory inside the OpenRAG container. Files added to the host or container directory are visible in both locations. To change this location, modify the **Documents Paths** variable in either the [**Advanced Setup** menu](/install#setup) or in the `.env` used by Docker Compose. diff --git a/docs/docs/get-started/docker.mdx b/docs/docs/get-started/docker.mdx index b9bb312a..af2e191f 100644 --- a/docs/docs/get-started/docker.mdx +++ b/docs/docs/get-started/docker.mdx @@ -18,7 +18,7 @@ OpenRAG has two Docker Compose files. Both files deploy the same applications an - Install the following: - - [Python](https://www.python.org/downloads/release/python-3100/) version 3.10 to 3.13. + - [Python](https://www.python.org/downloads/release/python-3100/) version 3.13 or later. - [uv](https://docs.astral.sh/uv/getting-started/installation/). - [Podman](https://podman.io/docs/installation) (recommended) or [Docker](https://docs.docker.com/get-docker/). - [`podman-compose`](https://docs.podman.io/en/latest/markdown/podman-compose.1.html) or [Docker Compose](https://docs.docker.com/compose/install/). To use Docker Compose with Podman, you must alias Docker Compose commands to Podman commands. @@ -187,7 +187,7 @@ docker compose up -d --force-recreate Reset state by rebuilding all of your containers. Your OpenSearch and Langflow databases will be lost. -Documents stored in the `./documents` directory will persist, since the directory is mounted as a volume in the OpenRAG backend container. +Documents stored in the `./openrag-documents` directory will persist, since the directory is mounted as a volume in the OpenRAG backend container. ```bash docker compose up --build --force-recreate --remove-orphans diff --git a/docs/docs/get-started/install.mdx b/docs/docs/get-started/install.mdx index 7e2c1e09..b30784b7 100644 --- a/docs/docs/get-started/install.mdx +++ b/docs/docs/get-started/install.mdx @@ -22,7 +22,7 @@ If you prefer running Podman or Docker containers and manually editing `.env` fi ## Prerequisites -- All OpenRAG installations require [Python](https://www.python.org/downloads/release/python-3100/) version 3.10 to 3.13. +- All OpenRAG installations require [Python](https://www.python.org/downloads/release/python-3100/) version 3.13 or later. - If you aren't using the automatic installer script, install the following: diff --git a/docs/docs/get-started/quickstart.mdx b/docs/docs/get-started/quickstart.mdx index 77f61b15..5a0e09c8 100644 --- a/docs/docs/get-started/quickstart.mdx +++ b/docs/docs/get-started/quickstart.mdx @@ -19,7 +19,7 @@ This quickstart requires the following: This quickstart uses OpenAI for simplicity. For other providers, see the complete [installation guide](/install). -- [Python](https://www.python.org/downloads/release/python-3100/) version 3.10 to 3.13. +- [Python](https://www.python.org/downloads/release/python-3100/) version 3.13 or later. - Microsoft Windows only: To run OpenRAG on Windows, you must use the Windows Subsystem for Linux (WSL). @@ -102,7 +102,7 @@ You can click a document to view the chunks of the document as they are stored i For this quickstart, use either the