ragflow/rag/llm
Dominik Novotný 1c84d1b562
Fix: azure OpenAI retry (#10213)
### What problem does this PR solve?

Currently, Azure OpenAI returns one minute Quota limit responses when
chat API is utilized. This change is needed in order to be able to
process almost any documents using models deployed in Azure Foundry.

### Type of change

- [x] Bug Fix (non-breaking change which fixes an issue)
2025-09-23 12:19:28 +08:00
..
__init__.py Feat: add support for Anthropic third-party API (#10173) 2025-09-19 19:06:14 +08:00
chat_model.py Fix: azure OpenAI retry (#10213) 2025-09-23 12:19:28 +08:00
cv_model.py Fix bug: broken import from rag.prompts.prompts (#10217) 2025-09-23 10:19:25 +08:00
embedding_model.py Refactor: use the same implement for total token count from res (#10197) 2025-09-22 17:17:06 +08:00
rerank_model.py Refactor: use the same implement for total token count from res (#10197) 2025-09-22 17:17:06 +08:00
sequence2txt_model.py Feat: add CometAPI to LLMFactory and update related mappings (#10119) 2025-09-18 09:51:29 +08:00
tts_model.py Feat: add CometAPI to LLMFactory and update related mappings (#10119) 2025-09-18 09:51:29 +08:00