* added telemetry utils
* added telemetry to openrag
* fixed http timeout
* Added OS and GPU logging
* Track task fail and cancel
* Updated messages to be more readable
* Changed backend to mount config at volume
* update lock
* Changed backend to reapply settings after detecting that flow is reset
* Added periodic backup for flows, make better reset
* tui warning
* Changed settings page to alert user that he has to disable lock flow
* Changed flows to be locked
* Do periodic backup only if onboarding is done
* Change backup function to only back up flows if flow lock is disabled
* Added session manager to reapply all settings
---------
Co-authored-by: Sebastián Estévez <estevezsebastian@gmail.com>
Introduces a 'fail_safe_mode' option to the Embedding Model and OpenSearch (Multi-Model Multi-Embedding) components, allowing errors to be logged and None returned instead of raising exceptions. Refactors embedding model fetching logic for better error handling and updates component metadata, field order, and dependencies. Also adds 'className' fields and updates frontend node folder IDs for improved UI consistency.
Replaces all references to 'OpenSearchHybrid-Ve6bS' with 'OpenSearchVectorStoreComponentMultimodalMultiEmbedding-By9U4' in main.py, processors, and file service. Adds a utility for injecting provider credentials into Langflow request headers and integrates it into chat and file services for improved credential handling.
Replaced imports from config_manager with settings in chat_service.py and langflow_file_service.py to use get_openrag_config from config.settings. This change ensures consistency with the updated configuration structure.
Introduces SELECTED_EMBEDDING_MODEL as a global environment variable in docker-compose files and ensures it is passed in API headers for Langflow-related services. Updates settings and onboarding logic to set this variable without triggering flow updates, improving embedding model configuration consistency across services.
Introduces fields and validation for Anthropic API key, Ollama endpoint, and IBM watsonx.ai API key, endpoint, and project ID in environment management and configuration screens. Updates validation utilities and config UI to support these providers, allowing users to set and validate credentials and endpoints for additional AI services.
* remove connection dot indicators on settings page, better toast message for provider setup dialogs, fix typo in default agent prompt
* format
* open llm model select when toast button to settings is clicked