* Changed backend to mount config at volume
* update lock
* Changed backend to reapply settings after detecting that flow is reset
* Added periodic backup for flows, make better reset
* tui warning
* Changed settings page to alert user that he has to disable lock flow
* Changed flows to be locked
* Do periodic backup only if onboarding is done
* Change backup function to only back up flows if flow lock is disabled
* Added session manager to reapply all settings
---------
Co-authored-by: Sebastián Estévez <estevezsebastian@gmail.com>
Uncommented build sections for backend, frontend, and langflow services in docker-compose.yml to allow local image builds. Updated SELECTED_EMBEDDING_MODEL to use environment variable in both docker-compose.yml and docker-compose-cpu.yml for improved configurability.
Introduces SELECTED_EMBEDDING_MODEL as a global environment variable in docker-compose files and ensures it is passed in API headers for Langflow-related services. Updates settings and onboarding logic to set this variable without triggering flow updates, improving embedding model configuration consistency across services.
Introduced new environment variables for Anthropic, WatsonX, and Ollama API integration in both backend and Langflow services within docker-compose files. Also enabled backend service to build from local Dockerfile instead of using only the image.
* fixed docker to old way
* added refetch when nudges are not fetched
* show skip overview just when finishing embedding
* just show children when show layout
* increased total onboarding steps
* made assistant message be saved on local storage to save progress on onboarding
* clean all items from local storage on finish onboarding
* only show navigation when onboarding is complete
* fixed navigation style to not placeholder for new conversation, not show loading state, and to show correct files
* fixed chat losing past message when navigating out of the chat
* fixed conversation be selected even though its a new conversation
* set the messages as just the user message when messages.length is 1
* fixed conversation be selected when exiting onboarding
* Added flows with new components
* commented model provider assignment
* Added agent component display name
* commented provider assignment, assign provider on the generic component, assign custom values
* fixed ollama not showing loading steps, fixed loading steps never being removed
* made embedding and llm model optional on onboarding call
* added isEmbedding handling on useModelSelection
* added isEmbedding on onboarding card, separating embedding from non embedding card
* Added one additional step to configure embeddings
* Added embedding provider config
* Changed settings.py to return if not embedding
* Added editing fields to onboarding
* updated onboarding and flows_service to change embedding and llm separately
* updated templates that needs to be changed with provider values
* updated flows with new components
* Changed config manager to not have default models
* Changed flows_service settings
* Complete steps if not embedding
* Add more onboarding steps
* Removed one step from llm steps
* Added Anthropic as a model for the language model on the frontend
* Added anthropic models
* Added anthropic support on Backend
* Fixed provider health and validation
* Format settings
* Change anthropic logo
* Changed button to not jump
* Changed flows service to make anthropic work
* Fixed some things
* add embedding specific global variables
* updated flows
* fixed ingestion flow
* Implemented anthropic on settings page
* add embedding provider logo
* updated backend to work with multiple provider config
* update useUpdateSettings with new settings type
* updated provider health banner to check for health with new api
* changed queries and mutations to use new api
* changed embedding model input to work with new api
* Implemented provider based config on the frontend
* update existing design
* fixed settings configured
* fixed provider health query to include health check for both the providers
* Changed model-providers to show correctly the configured providers
* Updated prompt
* updated openrag agent
* Fixed settings to allow editing providers and changing llm and embedding models
* updated settings
* changed lf ver
* bump openrag version
* added more steps
* update settings to create the global variables
* updated steps
* updated default prompt
---------
Co-authored-by: Sebastián Estévez <estevezsebastian@gmail.com>
Introduces the CONNECTOR_TYPE_URL environment variable to docker-compose files and assets, updates the OpenRAG URL ingestion flow to use it, and ensures it is set in the auth service global variables. This enables explicit configuration and handling of URL-based connectors in the OpenRAG system.
Added new environment variables (FILENAME, MIMETYPE, FILESIZE) and updated LANGFLOW_VARIABLES_TO_GET_FROM_ENVIRONMENT to include them. Comment formatting and container_name lines were adjusted for consistency, and DEFAULT_FOLDER_NAME is now commented out in both files.
* hard-coded openai models
* ensure index if disable ingest with langflow is active
* update backend to not update embedding model when flag is disabled
* initialize index on startup when feature flag is enabled
* put config.yaml on docker compose
This commit modifies both docker-compose.yml and docker-compose-cpu.yml to update the environment variable from FLOW_ID to LANGFLOW_CHAT_FLOW_ID, ensuring consistency across configurations. These changes contribute to a more robust and well-documented codebase.
This commit modifies both docker-compose.yml and docker-compose-cpu.yml to add OPENSEARCH_PASSWORD to the LANGFLOW_VARIABLES_TO_GET_FROM_ENVIRONMENT. Additionally, unnecessary whitespace has been removed for improved readability.