diff --git a/.github/workflows/build-langflow-responses.yml b/.github/workflows/build-langflow-responses.yml new file mode 100644 index 00000000..0f9d3d08 --- /dev/null +++ b/.github/workflows/build-langflow-responses.yml @@ -0,0 +1,59 @@ +name: Build Langflow Responses Multi-Arch + +on: + workflow_dispatch: + +jobs: + build: + strategy: + fail-fast: false + matrix: + include: + - platform: linux/amd64 + arch: amd64 + runs-on: ubuntu-latest + - platform: linux/arm64 + arch: arm64 + runs-on: [self-hosted, linux, ARM64, langflow-ai-arm64-2] + + runs-on: ${{ matrix.runs-on }} + + steps: + - name: Checkout + uses: actions/checkout@v4 + + - name: Set up Docker Buildx + uses: docker/setup-buildx-action@v3 + + - name: Login to Docker Hub + uses: docker/login-action@v3 + with: + username: ${{ secrets.DOCKER_USERNAME }} + password: ${{ secrets.DOCKER_PASSWORD }} + + - name: Build and push langflow (${{ matrix.arch }}) + uses: docker/build-push-action@v5 + with: + context: . + file: ./Dockerfile.langflow + platforms: ${{ matrix.platform }} + push: true + tags: phact/langflow:responses-${{ matrix.arch }} + cache-from: type=gha,scope=langflow-responses-${{ matrix.arch }} + cache-to: type=gha,mode=max,scope=langflow-responses-${{ matrix.arch }} + + manifest: + needs: build + runs-on: ubuntu-latest + steps: + - name: Login to Docker Hub + uses: docker/login-action@v3 + with: + username: ${{ secrets.DOCKER_USERNAME }} + password: ${{ secrets.DOCKER_PASSWORD }} + + - name: Create and push multi-arch manifest + run: | + docker buildx imagetools create -t phact/langflow:responses \ + phact/langflow:responses-amd64 \ + phact/langflow:responses-arm64 \ No newline at end of file diff --git a/.gitignore b/.gitignore index 4f22035a..8bf471e7 100644 --- a/.gitignore +++ b/.gitignore @@ -18,3 +18,5 @@ wheels/ 1001*.pdf *.json .DS_Store + +config.yaml diff --git a/config.yaml b/config.yaml deleted file mode 100644 index 3bafb8bd..00000000 --- a/config.yaml +++ /dev/null @@ -1,31 +0,0 @@ -# OpenRAG Configuration File -# This file allows you to configure OpenRAG settings. -# Environment variables will override these settings unless edited is true. - -# Track if this config has been manually edited (prevents env var overrides) -edited: false - -# Model provider configuration -provider: - # Supported providers: "openai", "anthropic", "azure", etc. - model_provider: "openai" - # API key for the model provider (can also be set via OPENAI_API_KEY env var) - api_key: "" - -# Knowledge base and document processing configuration -knowledge: - # Embedding model for vector search - embedding_model: "text-embedding-3-small" - # Text chunk size for document processing - chunk_size: 1000 - # Overlap between chunks - chunk_overlap: 200 - # Docling preset setting - doclingPresets: standard - -# AI agent configuration -agent: - # Language model for the chat agent - llm_model: "gpt-4o-mini" - # System prompt for the agent - system_prompt: "You are a helpful AI assistant with access to a knowledge base. Answer questions based on the provided context." diff --git a/docs/docs/core-components/knowledge.mdx b/docs/docs/core-components/knowledge.mdx index 255b0b68..2ea5ef9f 100644 --- a/docs/docs/core-components/knowledge.mdx +++ b/docs/docs/core-components/knowledge.mdx @@ -38,13 +38,9 @@ The file is loaded into your OpenSearch database, and appears in the Knowledge p To load and process a directory from the mapped location, click