ragflow/rag
Edouard Hur b29539b442 Fix: CoHereRerank not respecting base_url when provided (#5784)
### What problem does this PR solve?

vLLM provider with a reranking model does not work : as vLLM uses under
the hood the [CoHereRerank
provider](https://github.com/infiniflow/ragflow/blob/v0.17.0/rag/llm/__init__.py#L250)
with a `base_url`, if this URL [is not passed to the Cohere
client](https://github.com/infiniflow/ragflow/blob/v0.17.0/rag/llm/rerank_model.py#L379-L382)
any attempt will endup on the Cohere SaaS (sending your private api key
in the process) instead of your vLLM instance.

### Type of change

- [x] Bug Fix (non-breaking change which fixes an issue)
- [ ] New Feature (non-breaking change which adds functionality)
- [ ] Documentation Update
- [ ] Refactoring
- [ ] Performance Improvement
- [ ] Other (please describe):
2025-03-10 11:22:06 +08:00
..
app Refactor: Optimize error handling and support parsing of XLS(EXCEL97—2003) files. (#5633) 2025-03-05 11:55:27 +08:00
llm Fix: CoHereRerank not respecting base_url when provided (#5784) 2025-03-10 11:22:06 +08:00
nlp Fix: bad case for tokenizer. (#5543) 2025-03-03 15:36:16 +08:00
res
svr Refactored DocumentService.update_progress (#5642) 2025-03-05 14:48:03 +08:00
utils refactor: no need to inherit in python3 clean the code (#5659) 2025-03-05 18:03:53 +08:00
__init__.py Update comments (#4569) 2025-01-21 20:52:28 +08:00
benchmark.py Refactor embedding batch_size (#3825) 2024-12-03 16:22:39 +08:00
prompts.py Fix: truncate message issue. (#5776) 2025-03-07 17:41:56 +08:00
raptor.py Made task_executor async to speedup parsing (#5530) 2025-03-03 18:59:49 +08:00
settings.py Feat: Accessing Alibaba Cloud OSS with Amazon S3 SDK (#5438) 2025-02-27 17:02:42 +08:00