Commit graph

176 commits

Author SHA1 Message Date
phact
304e0c15d2 v1 endpoint 2025-12-19 03:36:02 -05:00
phact
bb71167d11 v1 endpoints for kf and settings 2025-12-19 03:31:54 -05:00
phact
320973a426 backend /v1 frontend /api/v1 routing, sdks port 3000 2025-12-19 01:50:29 -05:00
phact
9bc6f8b6eb ingest poll and other sdk endpoint fixes 2025-12-16 22:40:39 -05:00
phact
74cba85ae6 v0 sdks wip 2025-12-16 02:04:31 -05:00
Lucas Oliveira
6cc011d36e Added onboarding rollback on backend 2025-12-05 15:49:13 -03:00
Lucas Oliveira
1f528c2935 Show specific watsonX error on banner 2025-12-05 14:37:07 -03:00
Lucas Oliveira
8279cec761 wait for 2 seconds if its watsonx 2025-12-03 17:33:29 -03:00
Lucas Oliveira
9b08f1fcee create lightweight health check based on query param 2025-12-03 17:11:51 -03:00
Eric Hare
c5e09c47fe
Reduce timeout to 60 2025-12-02 14:11:30 -08:00
Eric Hare
133b59c2c6
Merge branch 'main' into feat-ollama-health-checks 2025-12-02 14:08:39 -08:00
Edwin Jose
c83ea54745
Merge branch 'main' into onboarding-fqs 2025-12-02 16:52:28 -05:00
Eric Hare
30ac7977f7
Merge branch 'main' into feat-ollama-health-checks 2025-12-02 13:32:27 -08:00
Eric Hare
d60432b1e0
Update provider_validation.py 2025-12-02 13:26:35 -08:00
Lucas Oliveira
ed77a9cba0
feat: add Scarf telemetry to OpenRAG (#548)
* added telemetry utils

* added telemetry to openrag

* fixed http timeout

* Added OS and GPU logging

* Track task fail and cancel

* Updated messages to be more readable
2025-12-02 16:52:51 -03:00
Eric Hare
a6f1ed7c48
feat: Provide ollama health checking 2025-12-02 11:16:34 -08:00
Sebastián Estévez
48d5e7c45e
Merge branch 'main' into onboarding-fqs 2025-12-02 13:15:31 -05:00
phact
0fc0be855c onboarding filter creation + sticky filters 2025-12-02 02:31:59 -05:00
phact
31c93846e6 fix fq ownership 2025-12-01 21:09:01 -05:00
phact
c106438d75 wip 2025-12-01 16:36:35 -05:00
Mike Fortman
55a685ebab fix get endpoints 2025-12-01 15:01:10 -06:00
Lucas Oliveira
cebb38eb04
fix: make backend reset langflow flows after upgrade, backup flows every 5 mins (#483)
* Changed backend to mount config at volume

* update lock

* Changed backend to reapply settings after detecting that flow is reset

* Added periodic backup for flows, make better reset

* tui warning

* Changed settings page to alert user that he has to disable lock flow

* Changed flows to be locked

* Do periodic backup only if onboarding is done

* Change backup function to only back up flows if flow lock is disabled

* Added session manager to reapply all settings

---------

Co-authored-by: Sebastián Estévez <estevezsebastian@gmail.com>
2025-11-26 22:53:33 -03:00
Edwin Jose
998d8b9195 Add fail-safe mode to embedding and OpenSearch components
Introduces a 'fail_safe_mode' option to the Embedding Model and OpenSearch (Multi-Model Multi-Embedding) components, allowing errors to be logged and None returned instead of raising exceptions. Refactors embedding model fetching logic for better error handling and updates component metadata, field order, and dependencies. Also adds 'className' fields and updates frontend node folder IDs for improved UI consistency.
2025-11-26 01:15:30 -05:00
Edwin Jose
284b982dbf Add SELECTED_EMBEDDING_MODEL global variable support
Introduces SELECTED_EMBEDDING_MODEL as a global environment variable in docker-compose files and ensures it is passed in API headers for Langflow-related services. Updates settings and onboarding logic to set this variable without triggering flow updates, improving embedding model configuration consistency across services.
2025-11-25 17:47:54 -05:00
Lucas Oliveira
c295431484
fix: refactor models validation to fix bugs related to ollama, watsonx and openai (#406)
* Fixed models service to try api key with first available model

* fixed ibm onboarding to not disable query when no data is available

* make ibm query disabled when not configured

* enable ollama query only when configured or endpoint present

* enable get openai models query when already configured

* just enable get from env when not configured

* Simplify ollama models validation

* fix max_tokens error on gpt 4o
2025-11-14 18:09:47 -03:00
Lucas Oliveira
37faf94979
feat: adds anthropic provider, splits onboarding editing into two, support provider changing with generic llm and embedding components (#373)
* Added flows with new components

* commented model provider assignment

* Added agent component display name

* commented provider assignment, assign provider on the generic component, assign custom values

* fixed ollama not showing loading steps, fixed loading steps never being removed

* made embedding and llm model optional on onboarding call

* added isEmbedding handling on useModelSelection

* added isEmbedding on onboarding card, separating embedding from non embedding card

* Added one additional step to configure embeddings

* Added embedding provider config

* Changed settings.py to return if not embedding

* Added editing fields to onboarding

* updated onboarding and flows_service to change embedding and llm separately

* updated templates that needs to be changed with provider values

* updated flows with new components

* Changed config manager to not have default models

* Changed flows_service settings

* Complete steps if not embedding

* Add more onboarding steps

* Removed one step from llm steps

* Added Anthropic as a model for the language model on the frontend

* Added anthropic models

* Added anthropic support on Backend

* Fixed provider health and validation

* Format settings

* Change anthropic logo

* Changed button to not jump

* Changed flows service to make anthropic work

* Fixed some things

* add embedding specific global variables

* updated flows

* fixed ingestion flow

* Implemented anthropic on settings page

* add embedding provider logo

* updated backend to work with multiple provider config

* update useUpdateSettings with new settings type

* updated provider health banner to check for health with new api

* changed queries and mutations to use new api

* changed embedding model input to work with new api

* Implemented provider based config on the frontend

* update existing design

* fixed settings configured

* fixed provider health query to include health check for both the providers

* Changed model-providers to show correctly the configured providers

* Updated prompt

* updated openrag agent

* Fixed settings to allow editing providers and changing llm and embedding models

* updated settings

* changed lf ver

* bump openrag version

* added more steps

* update settings to create the global variables

* updated steps

* updated default prompt

---------

Co-authored-by: Sebastián Estévez <estevezsebastian@gmail.com>
2025-11-11 19:22:16 -03:00
Lucas Oliveira
a5d25e0c0b
fix: disable upload message when ingesting on onboarding, wait for file to be ingested, added knowledge filters on nudges (#345)
* Removed upload start message

* Made onboarding upload refetch nudges and only finish when document is ingested

* Implemented query filters on nudges

* changed get to post

* Implemented filtering for documents that are not sample data on nudges

---------

Co-authored-by: Sebastián Estévez <estevezsebastian@gmail.com>
2025-11-11 18:20:39 -03:00
Cole Goldsmith
b88c8b20df
Feat/provider validation banner (#353)
* models query combined

* make endpoint to handle provider health

* provider health banner

* update-pdf-to-include-provider-selection (#344)

* polishing the error fixing experience

* fix agent instructions and up char limit

* fix provider

* disable tracing in langflow

* improve docling serve banner remove false positives

* Changed pyproject.toml docling versions

* Added another uv lock revision

* version bump

* unused things and fix bad conflicts

* add isFetching to the hook

* put back settings for models queries to never cache results

* update banner refetching indicator

* validate provider settings when saving

* fix settings page layout issue

* Added retry as false on the get models, to not take a long time

---------

Co-authored-by: Mendon Kissling <59585235+mendonk@users.noreply.github.com>
Co-authored-by: Mike Fortman <michael.fortman@datastax.com>
Co-authored-by: phact <estevezsebastian@gmail.com>
Co-authored-by: Lucas Oliveira <lucas.edu.oli@hotmail.com>
2025-11-06 13:03:50 -06:00
Mike Fortman
69d2132a33 fix provider 2025-11-04 12:59:23 -06:00
Cole Goldsmith
2d31c4b9b0
Feat/278 Edit current model provider settings (#307)
* update settings update api to allow changing model provider config

* use react hook form

* make settings page small width

* re-use the onboarding forms instead of rolling a custom one

* issue

* remove test

* make custom forms with react-hook-form

* replace the updateFlow mutation with updateSettings

* show all the model providers

* revert changes to onboarding forms

* disabled state styles for providers

* break model selectors into their own file

* use existing selector component, use settings endpoint instead of onboarding, clean up form styles

* revert changes to openai onboarding

* small form changes
2025-10-31 13:22:51 -05:00
Lucas Oliveira
b9ea9c99f1
fix: fixed bugs on ollama integration, added ingestion on onboarding (#330)
* Updated ollama components

* Changed ollama display name to be correct

* Changed prompt of provider validation

* removed event dispatched from file upload

* Changed onboarding to upload the entire knowledge

* Changed default models for ollama
2025-10-30 09:02:06 -03:00
Lucas Oliveira
7b635df9d0
fix: added better onboarding error handling, added probing api keys and models (#326)
* Added error showing to onboarding card

* Added error state on animated provider steps

* removed toast on error

* Fixed animation on onboarding card

* fixed animation time

* Implemented provider validation

* Added provider validation before ingestion

* Changed error border

* remove log

---------

Co-authored-by: Mike Fortman <michael.fortman@datastax.com>
2025-10-29 15:59:10 -03:00
phact
e3353bb0f8 loosen reconfigure check 2025-10-24 04:11:58 -04:00
Lucas Oliveira
cfd28ede6e Fix backend file context upload 2025-10-23 18:30:35 -03:00
phact
77edef26f7 fix conftest and more optionals 2025-10-14 12:17:07 -04:00
phact
9674021fae v0.1.24 2025-10-14 12:15:45 -04:00
Sebastián Estévez
b9f109ea7d
Merge branch 'main' into multi-embedding-support 2025-10-11 02:44:16 -04:00
phact
81901c666b fix: ensure /settings changes embedding models across all flows 2025-10-11 00:05:07 -04:00
Copilot
140d24603d
Implement dynamic Ollama embedding dimension resolution with server probing (#237)
* Initial plan

* Implement dynamic Ollama embedding dimension resolution with probing

Co-authored-by: phact <1313220+phact@users.noreply.github.com>

* Fix Ollama probing

* raise instead of dims 0

* Show better error

* Run embedding probe before saving settings so that user can update

---------

Co-authored-by: copilot-swe-agent[bot] <198982749+Copilot@users.noreply.github.com>
Co-authored-by: phact <1313220+phact@users.noreply.github.com>
Co-authored-by: Lucas Oliveira <lucas.edu.oli@hotmail.com>
Co-authored-by: phact <estevezsebastian@gmail.com>
2025-10-09 16:40:58 -03:00
phact
a5aab8e92a podman fixes: bind docling-serve to 0.0, improve logging, support magic podman and docker hostnames 2025-10-08 13:56:52 -04:00
phact
08f441c0e5 backend fix 2025-10-06 21:22:44 -04:00
phact
7763dc5d48 fix: file tuple for ingest 2025-10-06 16:25:05 -04:00
Lucas Oliveira
95bdc59af0 Use opensearch queries on documents and processors 2025-10-06 09:45:33 -03:00
Edwin Jose
5ea70cc779
Refactor file upload task to use filename mapping (#195)
Changed the handling of original filenames in Langflow upload tasks to use a mapping from file paths to original filenames instead of a list. Updated both the API router and TaskService to support this change, improving reliability when associating uploaded files with their original names.
2025-10-03 17:17:05 -03:00
Lucas Oliveira
589839fedc implement duplicate handling on backend and frontend 2025-10-03 15:34:54 -03:00
Lucas Oliveira
efb74f5ab0 Merge remote-tracking branch 'origin/main' into fix/tasks_ingest 2025-10-03 09:39:58 -03:00
Lucas Oliveira
179a7403cc fixed router, langflow files and processors to not use temp names 2025-10-02 18:09:18 -03:00
Lucas Oliveira
11464f031e Merge remote-tracking branch 'origin/main' into lfx-openrag-update-flows 2025-10-02 17:40:19 -03:00
Lucas Oliveira
52896a8da7
fix: set defined models for openai, make default embedding be used when disable ingest with LF feature flag is enabled (#128)
* hard-coded openai models

* ensure index if disable ingest with langflow is active

* update backend to not update embedding model when flag is disabled

* initialize index on startup when feature flag is enabled

* put config.yaml on docker compose
2025-10-01 16:50:21 -03:00
Lucas Oliveira
e4b9ebb186 deleted docling component id 2025-10-01 12:40:12 -03:00