ragflow/rag/llm
N0bodycan 9863862348
fix: prevent redundant retries in async_chat_streamly upon success (#11832)
## What changes were proposed in this pull request?
Added a return statement after the successful completion of the async
for loop in async_chat_streamly.

## Why are the changes needed?
Previously, the code lacked a break/return mechanism inside the try
block. This caused the retry loop (for attempt in range...) to continue
executing even after the LLM response was successfully generated and
yielded, resulting in duplicate requests (up to max_retries times).

## Does this PR introduce any user-facing change?
No (it fixes an internal logic bug).
2025-12-09 17:14:30 +08:00
..
__init__.py Refa: cleanup synchronous functions in chat_model and implement synchronization for conversation and dialog chats (#11779) 2025-12-08 09:43:03 +08:00
chat_model.py fix: prevent redundant retries in async_chat_streamly upon success (#11832) 2025-12-09 17:14:30 +08:00
cv_model.py Refa: migrate CV model chat to Async (#11828) 2025-12-09 13:08:37 +08:00
embedding_model.py Fix: jina embedding issue (#11628) 2025-12-01 14:24:35 +08:00
rerank_model.py fix cohere rerank base_url default (#11353) 2025-11-20 09:46:39 +08:00
sequence2txt_model.py Feat:new api /sequence2txt and update QWenSeq2txt (#11643) 2025-12-02 11:17:31 +08:00
tts_model.py Move token related functions to common (#10942) 2025-11-03 08:50:05 +08:00