Enable local build for OpenRAG services and update metadata handling
Switched OpenRAG backend and frontend in docker-compose.yml to use local Dockerfile builds instead of remote images. Updated environment variables for better clarity and system integration. In flows/openrag_agent.json and langflow_file_service, improved handling of docs_metadata to support Data objects and added logging for metadata ingestion. Added agent_llm edge to agent node in flow definition.
This commit is contained in:
parent
f4268bb99e
commit
392322077f
4 changed files with 67 additions and 262 deletions
|
|
@ -39,10 +39,10 @@ services:
|
||||||
- "5601:5601"
|
- "5601:5601"
|
||||||
|
|
||||||
openrag-backend:
|
openrag-backend:
|
||||||
image: phact/openrag-backend:${OPENRAG_VERSION:-latest}
|
# image: phact/openrag-backend:${OPENRAG_VERSION:-latest}
|
||||||
#build:
|
build:
|
||||||
#context: .
|
context: .
|
||||||
#dockerfile: Dockerfile.backend
|
dockerfile: Dockerfile.backend
|
||||||
container_name: openrag-backend
|
container_name: openrag-backend
|
||||||
depends_on:
|
depends_on:
|
||||||
- langflow
|
- langflow
|
||||||
|
|
@ -76,10 +76,11 @@ services:
|
||||||
gpus: all
|
gpus: all
|
||||||
|
|
||||||
openrag-frontend:
|
openrag-frontend:
|
||||||
image: phact/openrag-frontend:${OPENRAG_VERSION:-latest}
|
# image: phact/openrag-frontend:${OPENRAG_VERSION:-latest}
|
||||||
#build:
|
build:
|
||||||
#context: .
|
context: .
|
||||||
#dockerfile: Dockerfile.frontend
|
dockerfile: Dockerfile.frontend
|
||||||
|
#dockerfile: Dockerfile.frontend
|
||||||
container_name: openrag-frontend
|
container_name: openrag-frontend
|
||||||
depends_on:
|
depends_on:
|
||||||
- openrag-backend
|
- openrag-backend
|
||||||
|
|
@ -102,11 +103,11 @@ services:
|
||||||
- OPENAI_API_KEY=${OPENAI_API_KEY}
|
- OPENAI_API_KEY=${OPENAI_API_KEY}
|
||||||
- LANGFLOW_LOAD_FLOWS_PATH=/app/flows
|
- LANGFLOW_LOAD_FLOWS_PATH=/app/flows
|
||||||
- LANGFLOW_SECRET_KEY=${LANGFLOW_SECRET_KEY}
|
- LANGFLOW_SECRET_KEY=${LANGFLOW_SECRET_KEY}
|
||||||
- JWT="dummy"
|
- JWT=None
|
||||||
- OWNER="dummy"
|
- OWNER=None
|
||||||
- OWNER_NAME="dummy"
|
- OWNER_NAME=None
|
||||||
- OWNER_EMAIL="dummy"
|
- OWNER_EMAIL=None
|
||||||
- CONNECTOR_TYPE="dummy"
|
- CONNECTOR_TYPE=system
|
||||||
- OPENRAG-QUERY-FILTER="{}"
|
- OPENRAG-QUERY-FILTER="{}"
|
||||||
- OPENSEARCH_PASSWORD=${OPENSEARCH_PASSWORD}
|
- OPENSEARCH_PASSWORD=${OPENSEARCH_PASSWORD}
|
||||||
- LANGFLOW_VARIABLES_TO_GET_FROM_ENVIRONMENT=JWT,OPENRAG-QUERY-FILTER,OPENSEARCH_PASSWORD,OWNER,OWNER_NAME,OWNER_EMAIL,CONNECTOR_TYPE
|
- LANGFLOW_VARIABLES_TO_GET_FROM_ENVIRONMENT=JWT,OPENRAG-QUERY-FILTER,OPENSEARCH_PASSWORD,OWNER,OWNER_NAME,OWNER_EMAIL,CONNECTOR_TYPE
|
||||||
|
|
|
||||||
File diff suppressed because one or more lines are too long
|
|
@ -335,6 +335,10 @@ async def _ingest_default_documents_langflow(services, file_paths):
|
||||||
settings=None, # Use default ingestion settings
|
settings=None, # Use default ingestion settings
|
||||||
jwt_token=effective_jwt, # Use JWT token (anonymous if needed)
|
jwt_token=effective_jwt, # Use JWT token (anonymous if needed)
|
||||||
delete_after_ingest=True, # Clean up after ingestion
|
delete_after_ingest=True, # Clean up after ingestion
|
||||||
|
owner=None,
|
||||||
|
owner_name=anonymous_user.name,
|
||||||
|
owner_email=anonymous_user.email,
|
||||||
|
connector_type="system_default",
|
||||||
)
|
)
|
||||||
|
|
||||||
logger.info(
|
logger.info(
|
||||||
|
|
|
||||||
|
|
@ -98,7 +98,7 @@ class LangflowFileService:
|
||||||
|
|
||||||
# Pass metadata via tweaks to OpenSearch component
|
# Pass metadata via tweaks to OpenSearch component
|
||||||
metadata_tweaks = []
|
metadata_tweaks = []
|
||||||
if owner:
|
if owner or owner is None:
|
||||||
metadata_tweaks.append({"key": "owner", "value": owner})
|
metadata_tweaks.append({"key": "owner", "value": owner})
|
||||||
if owner_name:
|
if owner_name:
|
||||||
metadata_tweaks.append({"key": "owner_name", "value": owner_name})
|
metadata_tweaks.append({"key": "owner_name", "value": owner_name})
|
||||||
|
|
|
||||||
Loading…
Add table
Reference in a new issue