ragflow/rag/llm
Edouard Hur b29539b442 Fix: CoHereRerank not respecting base_url when provided (#5784)
### What problem does this PR solve?

vLLM provider with a reranking model does not work : as vLLM uses under
the hood the [CoHereRerank
provider](https://github.com/infiniflow/ragflow/blob/v0.17.0/rag/llm/__init__.py#L250)
with a `base_url`, if this URL [is not passed to the Cohere
client](https://github.com/infiniflow/ragflow/blob/v0.17.0/rag/llm/rerank_model.py#L379-L382)
any attempt will endup on the Cohere SaaS (sending your private api key
in the process) instead of your vLLM instance.

### Type of change

- [x] Bug Fix (non-breaking change which fixes an issue)
- [ ] New Feature (non-breaking change which adds functionality)
- [ ] Documentation Update
- [ ] Refactoring
- [ ] Performance Improvement
- [ ] Other (please describe):
2025-03-10 11:22:06 +08:00
..
__init__.py Feat: support huggingface re-rank model. (#5684) 2025-03-06 10:44:04 +08:00
chat_model.py Refa: remove max tokens since no one needs it. (#5690) 2025-03-06 11:29:40 +08:00
cv_model.py Fix: add image2text issue. (#5431) 2025-02-27 14:06:49 +08:00
embedding_model.py Add support to boto3 default connection (#5246) 2025-02-24 11:01:14 +08:00
rerank_model.py Fix: CoHereRerank not respecting base_url when provided (#5784) 2025-03-10 11:22:06 +08:00
sequence2txt_model.py Feat: add gpustack model provider (#4469) 2025-01-15 14:15:58 +08:00
tts_model.py Feat: add gpustack model provider (#4469) 2025-01-15 14:15:58 +08:00