From 917b078c20cb810565b114e30cb9af422f8ec8fd Mon Sep 17 00:00:00 2001 From: Mike Fortman Date: Tue, 23 Sep 2025 16:01:02 -0500 Subject: [PATCH] old tweaks cleanup --- frontend/src/app/settings/page.tsx | 2 ++ src/agent.py | 18 +----------------- src/api/settings.py | 16 +--------------- src/services/chat_service.py | 11 ----------- 4 files changed, 4 insertions(+), 43 deletions(-) diff --git a/frontend/src/app/settings/page.tsx b/frontend/src/app/settings/page.tsx index 50dc7867..eea555c2 100644 --- a/frontend/src/app/settings/page.tsx +++ b/frontend/src/app/settings/page.tsx @@ -152,6 +152,7 @@ function KnowledgeSourcesPage() { }, }); + // Debounced update function const debouncedUpdate = useDebounce( (variables: Parameters[0]) => { @@ -219,6 +220,7 @@ function KnowledgeSourcesPage() { // Update processing mode const handleProcessingModeChange = (mode: string) => { setProcessingMode(mode); + // Update the configuration setting (backend will also update the flow automatically) debouncedUpdate({ doclingPresets: mode }); }; diff --git a/src/agent.py b/src/agent.py index ab99f597..1eb362bd 100644 --- a/src/agent.py +++ b/src/agent.py @@ -106,7 +106,6 @@ async def async_response_stream( model: str, extra_headers: dict = None, previous_response_id: str = None, - tweaks: dict = None, log_prefix: str = "response", ): logger.info("User prompt received", prompt=prompt) @@ -121,8 +120,6 @@ async def async_response_stream( } if previous_response_id is not None: request_params["previous_response_id"] = previous_response_id - if tweaks: - request_params["tweaks"] = tweaks if "x-api-key" not in client.default_headers: if hasattr(client, "api_key") and extra_headers is not None: @@ -199,7 +196,6 @@ async def async_response( model: str, extra_headers: dict = None, previous_response_id: str = None, - tweaks: dict = None, log_prefix: str = "response", ): try: @@ -214,8 +210,6 @@ async def async_response( } if previous_response_id is not None: request_params["previous_response_id"] = previous_response_id - if tweaks: - request_params["tweaks"] = tweaks if extra_headers: request_params["extra_headers"] = extra_headers @@ -249,7 +243,6 @@ async def async_stream( model: str, extra_headers: dict = None, previous_response_id: str = None, - tweaks: dict = None, log_prefix: str = "response", ): async for chunk in async_response_stream( @@ -258,7 +251,6 @@ async def async_stream( model, extra_headers=extra_headers, previous_response_id=previous_response_id, - tweaks=tweaks, log_prefix=log_prefix, ): yield chunk @@ -271,7 +263,6 @@ async def async_langflow( prompt: str, extra_headers: dict = None, previous_response_id: str = None, - tweaks: dict = None, ): response_text, response_id, response_obj = await async_response( langflow_client, @@ -279,7 +270,6 @@ async def async_langflow( flow_id, extra_headers=extra_headers, previous_response_id=previous_response_id, - tweaks=tweaks, log_prefix="langflow", ) return response_text, response_id @@ -292,7 +282,6 @@ async def async_langflow_stream( prompt: str, extra_headers: dict = None, previous_response_id: str = None, - tweaks: dict = None, ): logger.debug("Starting langflow stream", prompt=prompt) try: @@ -302,8 +291,7 @@ async def async_langflow_stream( flow_id, extra_headers=extra_headers, previous_response_id=previous_response_id, - tweaks=tweaks, - log_prefix="langflow", + log_prefix="langflow", ): logger.debug( "Yielding chunk from langflow stream", @@ -463,7 +451,6 @@ async def async_langflow_chat( user_id: str, extra_headers: dict = None, previous_response_id: str = None, - tweaks: dict = None, store_conversation: bool = True, ): logger.debug( @@ -497,7 +484,6 @@ async def async_langflow_chat( flow_id, extra_headers=extra_headers, previous_response_id=previous_response_id, - tweaks=tweaks, log_prefix="langflow", ) logger.debug( @@ -576,7 +562,6 @@ async def async_langflow_chat_stream( user_id: str, extra_headers: dict = None, previous_response_id: str = None, - tweaks: dict = None, ): logger.debug( "async_langflow_chat_stream called", @@ -603,7 +588,6 @@ async def async_langflow_chat_stream( flow_id, extra_headers=extra_headers, previous_response_id=previous_response_id, - tweaks=tweaks, log_prefix="langflow", ): # Extract text content to build full response for history diff --git a/src/api/settings.py b/src/api/settings.py index 9723cdeb..560eb400 100644 --- a/src/api/settings.py +++ b/src/api/settings.py @@ -47,22 +47,7 @@ def get_docling_preset_configs(): } -def get_docling_tweaks(docling_preset: str = None) -> dict: - """Get Langflow tweaks for docling component based on preset""" - if not docling_preset: - # Get current preset from config - openrag_config = get_openrag_config() - docling_preset = openrag_config.knowledge.doclingPresets - preset_configs = get_docling_preset_configs() - - if docling_preset not in preset_configs: - docling_preset = "standard" # fallback - - preset_config = preset_configs[docling_preset] - docling_serve_opts = json.dumps(preset_config) - - return {"DoclingRemote-ayRdw": {"docling_serve_opts": docling_serve_opts}} async def get_settings(request, session_manager): @@ -626,3 +611,4 @@ async def update_docling_preset(request, session_manager): {"error": f"Failed to update docling preset: {str(e)}"}, status_code=500 ) + diff --git a/src/services/chat_service.py b/src/services/chat_service.py index 4b3c9d26..5ffe30f9 100644 --- a/src/services/chat_service.py +++ b/src/services/chat_service.py @@ -2,7 +2,6 @@ import json from config.settings import NUDGES_FLOW_ID, clients, LANGFLOW_URL, LANGFLOW_CHAT_FLOW_ID from agent import async_chat, async_langflow, async_chat_stream from auth_context import set_auth_context -from api.settings import get_docling_tweaks from utils.logging_config import get_logger logger = get_logger(__name__) @@ -127,8 +126,6 @@ class ChatService: "Langflow client not initialized. Ensure LANGFLOW is reachable or set LANGFLOW_KEY." ) - # Get docling tweaks based on current configuration - docling_tweaks = get_docling_tweaks() if stream: from agent import async_langflow_chat_stream @@ -140,7 +137,6 @@ class ChatService: user_id, extra_headers=extra_headers, previous_response_id=previous_response_id, - tweaks=docling_tweaks, ) else: from agent import async_langflow_chat @@ -152,7 +148,6 @@ class ChatService: user_id, extra_headers=extra_headers, previous_response_id=previous_response_id, - tweaks=docling_tweaks, ) response_data = {"response": response_text} if response_id: @@ -202,8 +197,6 @@ class ChatService: from agent import async_langflow_chat - # Get docling tweaks (might not be used by nudges flow, but keeping consistent) - docling_tweaks = get_docling_tweaks() response_text, response_id = await async_langflow_chat( langflow_client, @@ -211,7 +204,6 @@ class ChatService: prompt, user_id, extra_headers=extra_headers, - tweaks=docling_tweaks, store_conversation=False, ) response_data = {"response": response_text} @@ -242,8 +234,6 @@ class ChatService: raise ValueError( "Langflow client not initialized. Ensure LANGFLOW is reachable or set LANGFLOW_KEY." ) - # Get docling tweaks based on current configuration - docling_tweaks = get_docling_tweaks() response_text, response_id = await async_langflow( langflow_client=langflow_client, @@ -251,7 +241,6 @@ class ChatService: prompt=document_prompt, extra_headers=extra_headers, previous_response_id=previous_response_id, - tweaks=docling_tweaks, ) else: # chat # Set auth context for chat tools and provide user_id