ragflow/rag/llm
fansir efc4796f01
Fix ratelimit errors during document parsing (#6413)
### What problem does this PR solve?

When using the online large model API knowledge base to extract
knowledge graphs, frequent Rate Limit Errors were triggered,
causing document parsing to fail. This commit fixes the issue by
optimizing API calls in the following way:
Added exponential backoff and jitter to the API call to reduce the
frequency of Rate Limit Errors.


### Type of change

- [x] Bug Fix (non-breaking change which fixes an issue)
- [ ] New Feature (non-breaking change which adds functionality)
- [ ] Documentation Update
- [ ] Refactoring
- [ ] Performance Improvement
- [ ] Other (please describe):
2025-03-22 23:07:03 +08:00
..
__init__.py Feat: add TTS support for SILICONFLOW. (#6264) 2025-03-19 12:52:12 +08:00
chat_model.py Fix ratelimit errors during document parsing (#6413) 2025-03-22 23:07:03 +08:00
cv_model.py Feat: add vision LLM PDF parser (#6173) 2025-03-18 14:52:20 +08:00
embedding_model.py Fix: the error of Ollama embeddings interface returning "500 Internal Server Error" (#6350) 2025-03-21 15:25:48 +08:00
rerank_model.py Fix: huggingface rerank model issue. (#6385) 2025-03-21 12:43:32 +08:00
sequence2txt_model.py Feat: add gpustack model provider (#4469) 2025-01-15 14:15:58 +08:00
tts_model.py Feat: add TTS support for SILICONFLOW. (#6264) 2025-03-19 12:52:12 +08:00