Simplify Configuration

This commit is contained in:
zrguo 2025-07-08 11:16:34 +08:00
parent 75dd4f3498
commit f5c80d7cde
6 changed files with 210 additions and 201 deletions

View file

@ -2,24 +2,15 @@
This document explains how to configure and use the rerank functionality in LightRAG to improve retrieval quality. This document explains how to configure and use the rerank functionality in LightRAG to improve retrieval quality.
## ⚠️ Important: Parameter Priority
**QueryParam.top_k has higher priority than rerank_top_k configuration:**
- When you set `QueryParam(top_k=5)`, it will override the `rerank_top_k=10` setting in LightRAG configuration
- This means the actual number of documents sent to rerank will be determined by QueryParam.top_k
- For optimal rerank performance, always consider the top_k value in your QueryParam calls
- Example: `rag.aquery(query, param=QueryParam(mode="naive", top_k=20))` will use 20, not rerank_top_k
## Overview ## Overview
Reranking is an optional feature that improves the quality of retrieved documents by re-ordering them based on their relevance to the query. This is particularly useful when you want higher precision in document retrieval across all query modes (naive, local, global, hybrid, mix). Reranking is an optional feature that improves the quality of retrieved documents by re-ordering them based on their relevance to the query. This is particularly useful when you want higher precision in document retrieval across all query modes (naive, local, global, hybrid, mix).
## Architecture ## Architecture
The rerank integration follows the same design pattern as the LLM integration: The rerank integration follows a simplified design pattern:
- **Configurable Models**: Support for multiple rerank providers through a generic API - **Single Function Configuration**: All rerank settings (model, API keys, top_k, etc.) are contained within the rerank function
- **Async Processing**: Non-blocking rerank operations - **Async Processing**: Non-blocking rerank operations
- **Error Handling**: Graceful fallback to original results - **Error Handling**: Graceful fallback to original results
- **Optional Feature**: Can be enabled/disabled via configuration - **Optional Feature**: Can be enabled/disabled via configuration
@ -29,24 +20,11 @@ The rerank integration follows the same design pattern as the LLM integration:
### Environment Variables ### Environment Variables
Set these variables in your `.env` file or environment: Set this variable in your `.env` file or environment:
```bash ```bash
# Enable/disable reranking # Enable/disable reranking
ENABLE_RERANK=True ENABLE_RERANK=True
# Rerank model configuration
RERANK_MODEL=BAAI/bge-reranker-v2-m3
RERANK_MAX_ASYNC=4
RERANK_TOP_K=10
# API configuration
RERANK_API_KEY=your_rerank_api_key_here
RERANK_BASE_URL=https://api.your-provider.com/v1/rerank
# Provider-specific keys (optional alternatives)
JINA_API_KEY=your_jina_api_key_here
COHERE_API_KEY=your_cohere_api_key_here
``` ```
### Programmatic Configuration ### Programmatic Configuration
@ -55,15 +33,27 @@ COHERE_API_KEY=your_cohere_api_key_here
from lightrag import LightRAG from lightrag import LightRAG
from lightrag.rerank import custom_rerank, RerankModel from lightrag.rerank import custom_rerank, RerankModel
# Method 1: Using environment variables (recommended) # Method 1: Using a custom rerank function with all settings included
async def my_rerank_func(query: str, documents: list, top_k: int = None, **kwargs):
return await custom_rerank(
query=query,
documents=documents,
model="BAAI/bge-reranker-v2-m3",
base_url="https://api.your-provider.com/v1/rerank",
api_key="your_api_key_here",
top_k=top_k or 10, # Handle top_k within the function
**kwargs
)
rag = LightRAG( rag = LightRAG(
working_dir="./rag_storage", working_dir="./rag_storage",
llm_model_func=your_llm_func, llm_model_func=your_llm_func,
embedding_func=your_embedding_func, embedding_func=your_embedding_func,
# Rerank automatically configured from environment variables enable_rerank=True,
rerank_model_func=my_rerank_func,
) )
# Method 2: Explicit configuration # Method 2: Using RerankModel wrapper
rerank_model = RerankModel( rerank_model = RerankModel(
rerank_func=custom_rerank, rerank_func=custom_rerank,
kwargs={ kwargs={
@ -79,7 +69,6 @@ rag = LightRAG(
embedding_func=your_embedding_func, embedding_func=your_embedding_func,
enable_rerank=True, enable_rerank=True,
rerank_model_func=rerank_model.rerank, rerank_model_func=rerank_model.rerank,
rerank_top_k=10,
) )
``` ```
@ -112,7 +101,8 @@ result = await jina_rerank(
query="your query", query="your query",
documents=documents, documents=documents,
model="BAAI/bge-reranker-v2-m3", model="BAAI/bge-reranker-v2-m3",
api_key="your_jina_api_key" api_key="your_jina_api_key",
top_k=10
) )
``` ```
@ -125,7 +115,8 @@ result = await cohere_rerank(
query="your query", query="your query",
documents=documents, documents=documents,
model="rerank-english-v2.0", model="rerank-english-v2.0",
api_key="your_cohere_api_key" api_key="your_cohere_api_key",
top_k=10
) )
``` ```
@ -143,11 +134,7 @@ Reranking is automatically applied at these key retrieval stages:
| Parameter | Type | Default | Description | | Parameter | Type | Default | Description |
|-----------|------|---------|-------------| |-----------|------|---------|-------------|
| `enable_rerank` | bool | False | Enable/disable reranking | | `enable_rerank` | bool | False | Enable/disable reranking |
| `rerank_model_name` | str | "BAAI/bge-reranker-v2-m3" | Model identifier | | `rerank_model_func` | callable | None | Custom rerank function containing all configurations (model, API keys, top_k, etc.) |
| `rerank_model_max_async` | int | 4 | Max concurrent rerank calls |
| `rerank_top_k` | int | 10 | Number of top results to return ⚠️ **Overridden by QueryParam.top_k** |
| `rerank_model_func` | callable | None | Custom rerank function |
| `rerank_model_kwargs` | dict | {} | Additional rerank parameters |
## Example Usage ## Example Usage
@ -157,6 +144,18 @@ Reranking is automatically applied at these key retrieval stages:
import asyncio import asyncio
from lightrag import LightRAG, QueryParam from lightrag import LightRAG, QueryParam
from lightrag.llm.openai import gpt_4o_mini_complete, openai_embedding from lightrag.llm.openai import gpt_4o_mini_complete, openai_embedding
from lightrag.rerank import jina_rerank
async def my_rerank_func(query: str, documents: list, top_k: int = None, **kwargs):
"""Custom rerank function with all settings included"""
return await jina_rerank(
query=query,
documents=documents,
model="BAAI/bge-reranker-v2-m3",
api_key="your_jina_api_key_here",
top_k=top_k or 10, # Default top_k if not provided
**kwargs
)
async def main(): async def main():
# Initialize with rerank enabled # Initialize with rerank enabled
@ -165,6 +164,7 @@ async def main():
llm_model_func=gpt_4o_mini_complete, llm_model_func=gpt_4o_mini_complete,
embedding_func=openai_embedding, embedding_func=openai_embedding,
enable_rerank=True, enable_rerank=True,
rerank_model_func=my_rerank_func,
) )
# Insert documents # Insert documents
@ -176,7 +176,7 @@ async def main():
# Query with rerank (automatically applied) # Query with rerank (automatically applied)
result = await rag.aquery( result = await rag.aquery(
"Your question here", "Your question here",
param=QueryParam(mode="hybrid", top_k=5) # ⚠️ This top_k=5 overrides rerank_top_k param=QueryParam(mode="hybrid", top_k=5) # This top_k is passed to rerank function
) )
print(result) print(result)
@ -211,19 +211,19 @@ async def test_rerank():
## Best Practices ## Best Practices
1. **Parameter Priority Awareness**: Remember that QueryParam.top_k always overrides rerank_top_k configuration 1. **Self-Contained Functions**: Include all necessary configurations (API keys, models, top_k handling) within your rerank function
2. **Performance**: Use reranking selectively for better performance vs. quality tradeoff 2. **Performance**: Use reranking selectively for better performance vs. quality tradeoff
3. **API Limits**: Monitor API usage and implement rate limiting if needed 3. **API Limits**: Monitor API usage and implement rate limiting within your rerank function
4. **Fallback**: Always handle rerank failures gracefully (returns original results) 4. **Fallback**: Always handle rerank failures gracefully (returns original results)
5. **Top-k Selection**: Choose appropriate `top_k` values in QueryParam based on your use case 5. **Top-k Handling**: Handle top_k parameter appropriately within your rerank function
6. **Cost Management**: Consider rerank API costs in your budget planning 6. **Cost Management**: Consider rerank API costs in your budget planning
## Troubleshooting ## Troubleshooting
### Common Issues ### Common Issues
1. **API Key Missing**: Ensure `RERANK_API_KEY` or provider-specific keys are set 1. **API Key Missing**: Ensure API keys are properly configured within your rerank function
2. **Network Issues**: Check `RERANK_BASE_URL` and network connectivity 2. **Network Issues**: Check API endpoints and network connectivity
3. **Model Errors**: Verify the rerank model name is supported by your API 3. **Model Errors**: Verify the rerank model name is supported by your API
4. **Document Format**: Ensure documents have `content` or `text` fields 4. **Document Format**: Ensure documents have `content` or `text` fields

View file

@ -182,11 +182,3 @@ REDIS_URI=redis://localhost:6379
# Rerank Configuration # Rerank Configuration
ENABLE_RERANK=False ENABLE_RERANK=False
RERANK_MODEL=BAAI/bge-reranker-v2-m3
RERANK_MAX_ASYNC=4
RERANK_TOP_K=10
# Note: QueryParam.top_k in your code will override RERANK_TOP_K setting
# Rerank API Configuration
RERANK_API_KEY=your_rerank_api_key_here
RERANK_BASE_URL=https://api.your-provider.com/v1/rerank

View file

@ -4,19 +4,12 @@ LightRAG Rerank Integration Example
This example demonstrates how to use rerank functionality with LightRAG This example demonstrates how to use rerank functionality with LightRAG
to improve retrieval quality across different query modes. to improve retrieval quality across different query modes.
IMPORTANT: Parameter Priority
- QueryParam(top_k=N) has higher priority than rerank_top_k in LightRAG configuration
- If you set QueryParam(top_k=5), it will override rerank_top_k setting
- For optimal rerank performance, use appropriate top_k values in QueryParam
Configuration Required: Configuration Required:
1. Set your LLM API key and base URL in llm_model_func() 1. Set your LLM API key and base URL in llm_model_func()
2. Set your embedding API key and base URL in embedding_func() 2. Set your embedding API key and base URL in embedding_func()
3. Set your rerank API key and base URL in the rerank configuration 3. Set your rerank API key and base URL in the rerank configuration
4. Or use environment variables (.env file): 4. Or use environment variables (.env file):
- RERANK_API_KEY=your_actual_rerank_api_key - ENABLE_RERANK=True
- RERANK_BASE_URL=https://your-actual-rerank-endpoint/v1/rerank
- RERANK_MODEL=your_rerank_model_name
""" """
import asyncio import asyncio
@ -35,6 +28,7 @@ setup_logger("test_rerank")
if not os.path.exists(WORKING_DIR): if not os.path.exists(WORKING_DIR):
os.mkdir(WORKING_DIR) os.mkdir(WORKING_DIR)
async def llm_model_func( async def llm_model_func(
prompt, system_prompt=None, history_messages=[], **kwargs prompt, system_prompt=None, history_messages=[], **kwargs
) -> str: ) -> str:
@ -48,6 +42,7 @@ async def llm_model_func(
**kwargs, **kwargs,
) )
async def embedding_func(texts: list[str]) -> np.ndarray: async def embedding_func(texts: list[str]) -> np.ndarray:
return await openai_embed( return await openai_embed(
texts, texts,
@ -56,6 +51,20 @@ async def embedding_func(texts: list[str]) -> np.ndarray:
base_url="https://api.your-embedding-provider.com/v1", base_url="https://api.your-embedding-provider.com/v1",
) )
async def my_rerank_func(query: str, documents: list, top_k: int = None, **kwargs):
"""Custom rerank function with all settings included"""
return await custom_rerank(
query=query,
documents=documents,
model="BAAI/bge-reranker-v2-m3",
base_url="https://api.your-rerank-provider.com/v1/rerank",
api_key="your_rerank_api_key_here",
top_k=top_k or 10, # Default top_k if not provided
**kwargs,
)
async def create_rag_with_rerank(): async def create_rag_with_rerank():
"""Create LightRAG instance with rerank configuration""" """Create LightRAG instance with rerank configuration"""
@ -64,17 +73,7 @@ async def create_rag_with_rerank():
embedding_dim = test_embedding.shape[1] embedding_dim = test_embedding.shape[1]
print(f"Detected embedding dimension: {embedding_dim}") print(f"Detected embedding dimension: {embedding_dim}")
# Create rerank model # Method 1: Using custom rerank function
rerank_model = RerankModel(
rerank_func=custom_rerank,
kwargs={
"model": "BAAI/bge-reranker-v2-m3",
"base_url": "https://api.your-rerank-provider.com/v1/rerank",
"api_key": "your_rerank_api_key_here",
}
)
# Initialize LightRAG with rerank
rag = LightRAG( rag = LightRAG(
working_dir=WORKING_DIR, working_dir=WORKING_DIR,
llm_model_func=llm_model_func, llm_model_func=llm_model_func,
@ -83,17 +82,50 @@ async def create_rag_with_rerank():
max_token_size=8192, max_token_size=8192,
func=embedding_func, func=embedding_func,
), ),
# Rerank Configuration # Simplified Rerank Configuration
enable_rerank=True, enable_rerank=True,
rerank_model_func=rerank_model.rerank, rerank_model_func=my_rerank_func,
rerank_top_k=10, # Note: QueryParam.top_k will override this
) )
return rag return rag
async def create_rag_with_rerank_model():
"""Alternative: Create LightRAG instance using RerankModel wrapper"""
# Get embedding dimension
test_embedding = await embedding_func(["test"])
embedding_dim = test_embedding.shape[1]
print(f"Detected embedding dimension: {embedding_dim}")
# Method 2: Using RerankModel wrapper
rerank_model = RerankModel(
rerank_func=custom_rerank,
kwargs={
"model": "BAAI/bge-reranker-v2-m3",
"base_url": "https://api.your-rerank-provider.com/v1/rerank",
"api_key": "your_rerank_api_key_here",
},
)
rag = LightRAG(
working_dir=WORKING_DIR,
llm_model_func=llm_model_func,
embedding_func=EmbeddingFunc(
embedding_dim=embedding_dim,
max_token_size=8192,
func=embedding_func,
),
enable_rerank=True,
rerank_model_func=rerank_model.rerank,
)
return rag
async def test_rerank_with_different_topk(): async def test_rerank_with_different_topk():
""" """
Test rerank functionality with different top_k settings to demonstrate parameter priority Test rerank functionality with different top_k settings
""" """
print("🚀 Setting up LightRAG with Rerank functionality...") print("🚀 Setting up LightRAG with Rerank functionality...")
@ -105,7 +137,7 @@ async def test_rerank_with_different_topk():
"LightRAG is a powerful retrieval-augmented generation system with multiple query modes.", "LightRAG is a powerful retrieval-augmented generation system with multiple query modes.",
"Vector databases enable efficient similarity search in high-dimensional embedding spaces.", "Vector databases enable efficient similarity search in high-dimensional embedding spaces.",
"Natural language processing has evolved with large language models and transformers.", "Natural language processing has evolved with large language models and transformers.",
"Machine learning algorithms can learn patterns from data without explicit programming." "Machine learning algorithms can learn patterns from data without explicit programming.",
] ]
print("📄 Inserting sample documents...") print("📄 Inserting sample documents...")
@ -119,16 +151,14 @@ async def test_rerank_with_different_topk():
top_k_values = [2, 5, 10] top_k_values = [2, 5, 10]
for top_k in top_k_values: for top_k in top_k_values:
print(f"\n📊 Testing with QueryParam(top_k={top_k}) - overrides rerank_top_k=10:") print(f"\n📊 Testing with QueryParam(top_k={top_k}):")
# Test naive mode with specific top_k # Test naive mode with specific top_k
result = await rag.aquery( result = await rag.aquery(query, param=QueryParam(mode="naive", top_k=top_k))
query,
param=QueryParam(mode="naive", top_k=top_k)
)
print(f" Result length: {len(result)} characters") print(f" Result length: {len(result)} characters")
print(f" Preview: {result[:100]}...") print(f" Preview: {result[:100]}...")
async def test_direct_rerank(): async def test_direct_rerank():
"""Test rerank function directly""" """Test rerank function directly"""
print("\n🔧 Direct Rerank API Test") print("\n🔧 Direct Rerank API Test")
@ -139,7 +169,7 @@ async def test_direct_rerank():
{"content": "LightRAG supports advanced reranking capabilities"}, {"content": "LightRAG supports advanced reranking capabilities"},
{"content": "Vector search finds semantically similar documents"}, {"content": "Vector search finds semantically similar documents"},
{"content": "Natural language processing with modern transformers"}, {"content": "Natural language processing with modern transformers"},
{"content": "The quick brown fox jumps over the lazy dog"} {"content": "The quick brown fox jumps over the lazy dog"},
] ]
query = "rerank improve quality" query = "rerank improve quality"
@ -153,7 +183,7 @@ async def test_direct_rerank():
model="BAAI/bge-reranker-v2-m3", model="BAAI/bge-reranker-v2-m3",
base_url="https://api.your-rerank-provider.com/v1/rerank", base_url="https://api.your-rerank-provider.com/v1/rerank",
api_key="your_rerank_api_key_here", api_key="your_rerank_api_key_here",
top_k=3 top_k=3,
) )
print("\n✅ Rerank Results:") print("\n✅ Rerank Results:")
@ -165,6 +195,7 @@ async def test_direct_rerank():
except Exception as e: except Exception as e:
print(f"❌ Rerank failed: {e}") print(f"❌ Rerank failed: {e}")
async def main(): async def main():
"""Main example function""" """Main example function"""
print("🎯 LightRAG Rerank Integration Example") print("🎯 LightRAG Rerank Integration Example")
@ -179,15 +210,17 @@ async def main():
print("\n✅ Example completed successfully!") print("\n✅ Example completed successfully!")
print("\n💡 Key Points:") print("\n💡 Key Points:")
print("QueryParam.top_k has higher priority than rerank_top_k") print("All rerank configurations are contained within rerank_model_func")
print(" ✓ Rerank improves document relevance ordering") print(" ✓ Rerank improves document relevance ordering")
print(" ✓ Configure API keys in your .env file for production") print(" ✓ Configure API keys within your rerank function")
print(" ✓ Monitor API usage and costs when using rerank services") print(" ✓ Monitor API usage and costs when using rerank services")
except Exception as e: except Exception as e:
print(f"\n❌ Example failed: {e}") print(f"\n❌ Example failed: {e}")
import traceback import traceback
traceback.print_exc() traceback.print_exc()
if __name__ == "__main__": if __name__ == "__main__":
asyncio.run(main()) asyncio.run(main())

View file

@ -249,25 +249,7 @@ class LightRAG:
"""Enable reranking for improved retrieval quality. Defaults to False.""" """Enable reranking for improved retrieval quality. Defaults to False."""
rerank_model_func: Callable[..., object] | None = field(default=None) rerank_model_func: Callable[..., object] | None = field(default=None)
"""Function for reranking retrieved documents. Optional.""" """Function for reranking retrieved documents. All rerank configurations (model name, API keys, top_k, etc.) should be included in this function. Optional."""
rerank_model_name: str = field(
default=os.getenv("RERANK_MODEL", "BAAI/bge-reranker-v2-m3")
)
"""Name of the rerank model used for reranking documents."""
rerank_model_max_async: int = field(default=int(os.getenv("RERANK_MAX_ASYNC", 4)))
"""Maximum number of concurrent rerank calls."""
rerank_model_kwargs: dict[str, Any] = field(default_factory=dict)
"""Additional keyword arguments passed to the rerank model function."""
rerank_top_k: int = field(default=int(os.getenv("RERANK_TOP_K", 10)))
"""Number of top documents to return after reranking.
Note: This value will be overridden by QueryParam.top_k in query calls.
Example: QueryParam(top_k=5) will override rerank_top_k=10 setting.
"""
# Storage # Storage
# --- # ---
@ -475,14 +457,6 @@ class LightRAG:
# Init Rerank # Init Rerank
if self.enable_rerank and self.rerank_model_func: if self.enable_rerank and self.rerank_model_func:
self.rerank_model_func = priority_limit_async_func_call(
self.rerank_model_max_async
)(
partial(
self.rerank_model_func, # type: ignore
**self.rerank_model_kwargs,
)
)
logger.info("Rerank model initialized for improved retrieval quality") logger.info("Rerank model initialized for improved retrieval quality")
elif self.enable_rerank and not self.rerank_model_func: elif self.enable_rerank and not self.rerank_model_func:
logger.warning( logger.warning(

View file

@ -2864,19 +2864,15 @@ async def apply_rerank_if_enabled(
return retrieved_docs return retrieved_docs
try: try:
# Determine top_k for reranking
rerank_top_k = top_k or global_config.get("rerank_top_k", 10)
rerank_top_k = min(rerank_top_k, len(retrieved_docs))
logger.debug( logger.debug(
f"Applying rerank to {len(retrieved_docs)} documents, returning top {rerank_top_k}" f"Applying rerank to {len(retrieved_docs)} documents, returning top {top_k}"
) )
# Apply reranking # Apply reranking - let rerank_model_func handle top_k internally
reranked_docs = await rerank_func( reranked_docs = await rerank_func(
query=query, query=query,
documents=retrieved_docs, documents=retrieved_docs,
top_k=rerank_top_k, top_k=top_k,
) )
if reranked_docs and len(reranked_docs) > 0: if reranked_docs and len(reranked_docs) > 0:
@ -2886,7 +2882,7 @@ async def apply_rerank_if_enabled(
return reranked_docs return reranked_docs
else: else:
logger.warning("Rerank returned empty results, using original documents") logger.warning("Rerank returned empty results, using original documents")
return retrieved_docs[:rerank_top_k] if rerank_top_k else retrieved_docs return retrieved_docs
except Exception as e: except Exception as e:
logger.error(f"Error during reranking: {e}, using original documents") logger.error(f"Error during reranking: {e}, using original documents")

View file

@ -1,12 +1,9 @@
from __future__ import annotations from __future__ import annotations
import os import os
import json
import aiohttp import aiohttp
import numpy as np
from typing import Callable, Any, List, Dict, Optional from typing import Callable, Any, List, Dict, Optional
from pydantic import BaseModel, Field from pydantic import BaseModel, Field
from dataclasses import asdict
from .utils import logger from .utils import logger
@ -15,14 +12,17 @@ class RerankModel(BaseModel):
""" """
Pydantic model class for defining a custom rerank model. Pydantic model class for defining a custom rerank model.
This class provides a convenient wrapper for rerank functions, allowing you to
encapsulate all rerank configurations (API keys, model settings, etc.) in one place.
Attributes: Attributes:
rerank_func (Callable[[Any], List[Dict]]): A callable function that reranks documents. rerank_func (Callable[[Any], List[Dict]]): A callable function that reranks documents.
The function should take query and documents as input and return reranked results. The function should take query and documents as input and return reranked results.
kwargs (Dict[str, Any]): A dictionary that contains the arguments to pass to the callable function. kwargs (Dict[str, Any]): A dictionary that contains the arguments to pass to the callable function.
This could include parameters such as the model name, API key, etc. This should include all necessary configurations such as model name, API key, base_url, etc.
Example usage: Example usage:
Rerank model example from jina: Rerank model example with Jina:
```python ```python
rerank_model = RerankModel( rerank_model = RerankModel(
rerank_func=jina_rerank, rerank_func=jina_rerank,
@ -32,6 +32,32 @@ class RerankModel(BaseModel):
"base_url": "https://api.jina.ai/v1/rerank" "base_url": "https://api.jina.ai/v1/rerank"
} }
) )
# Use in LightRAG
rag = LightRAG(
enable_rerank=True,
rerank_model_func=rerank_model.rerank,
# ... other configurations
)
```
Or define a custom function directly:
```python
async def my_rerank_func(query: str, documents: list, top_k: int = None, **kwargs):
return await jina_rerank(
query=query,
documents=documents,
model="BAAI/bge-reranker-v2-m3",
api_key="your_api_key_here",
top_k=top_k or 10,
**kwargs
)
rag = LightRAG(
enable_rerank=True,
rerank_model_func=my_rerank_func,
# ... other configurations
)
``` ```
""" """
@ -43,16 +69,13 @@ class RerankModel(BaseModel):
query: str, query: str,
documents: List[Dict[str, Any]], documents: List[Dict[str, Any]],
top_k: Optional[int] = None, top_k: Optional[int] = None,
**extra_kwargs **extra_kwargs,
) -> List[Dict[str, Any]]: ) -> List[Dict[str, Any]]:
"""Rerank documents using the configured model function.""" """Rerank documents using the configured model function."""
# Merge extra kwargs with model kwargs # Merge extra kwargs with model kwargs
kwargs = {**self.kwargs, **extra_kwargs} kwargs = {**self.kwargs, **extra_kwargs}
return await self.rerank_func( return await self.rerank_func(
query=query, query=query, documents=documents, top_k=top_k, **kwargs
documents=documents,
top_k=top_k,
**kwargs
) )
@ -73,7 +96,7 @@ class MultiRerankModel(BaseModel):
documents: List[Dict[str, Any]], documents: List[Dict[str, Any]],
mode: str = "default", mode: str = "default",
top_k: Optional[int] = None, top_k: Optional[int] = None,
**kwargs **kwargs,
) -> List[Dict[str, Any]]: ) -> List[Dict[str, Any]]:
"""Rerank using the appropriate model based on mode.""" """Rerank using the appropriate model based on mode."""
@ -100,7 +123,7 @@ async def generic_rerank_api(
base_url: str, base_url: str,
api_key: str, api_key: str,
top_k: Optional[int] = None, top_k: Optional[int] = None,
**kwargs **kwargs,
) -> List[Dict[str, Any]]: ) -> List[Dict[str, Any]]:
""" """
Generic rerank function that works with Jina/Cohere compatible APIs. Generic rerank function that works with Jina/Cohere compatible APIs.
@ -129,23 +152,15 @@ async def generic_rerank_api(
for doc in documents: for doc in documents:
if isinstance(doc, dict): if isinstance(doc, dict):
# Use 'content' field if available, otherwise use 'text' or convert to string # Use 'content' field if available, otherwise use 'text' or convert to string
text = doc.get('content') or doc.get('text') or str(doc) text = doc.get("content") or doc.get("text") or str(doc)
else: else:
text = str(doc) text = str(doc)
prepared_docs.append(text) prepared_docs.append(text)
# Prepare request # Prepare request
headers = { headers = {"Content-Type": "application/json", "Authorization": f"Bearer {api_key}"}
"Content-Type": "application/json",
"Authorization": f"Bearer {api_key}"
}
data = { data = {"model": model, "query": query, "documents": prepared_docs, **kwargs}
"model": model,
"query": query,
"documents": prepared_docs,
**kwargs
}
if top_k is not None: if top_k is not None:
data["top_k"] = min(top_k, len(prepared_docs)) data["top_k"] = min(top_k, len(prepared_docs))
@ -170,7 +185,9 @@ async def generic_rerank_api(
if 0 <= doc_idx < len(documents): if 0 <= doc_idx < len(documents):
reranked_doc = documents[doc_idx].copy() reranked_doc = documents[doc_idx].copy()
if "relevance_score" in item: if "relevance_score" in item:
reranked_doc["rerank_score"] = item["relevance_score"] reranked_doc["rerank_score"] = item[
"relevance_score"
]
reranked_docs.append(reranked_doc) reranked_docs.append(reranked_doc)
return reranked_docs return reranked_docs
else: else:
@ -189,7 +206,7 @@ async def jina_rerank(
top_k: Optional[int] = None, top_k: Optional[int] = None,
base_url: str = "https://api.jina.ai/v1/rerank", base_url: str = "https://api.jina.ai/v1/rerank",
api_key: Optional[str] = None, api_key: Optional[str] = None,
**kwargs **kwargs,
) -> List[Dict[str, Any]]: ) -> List[Dict[str, Any]]:
""" """
Rerank documents using Jina AI API. Rerank documents using Jina AI API.
@ -216,7 +233,7 @@ async def jina_rerank(
base_url=base_url, base_url=base_url,
api_key=api_key, api_key=api_key,
top_k=top_k, top_k=top_k,
**kwargs **kwargs,
) )
@ -227,7 +244,7 @@ async def cohere_rerank(
top_k: Optional[int] = None, top_k: Optional[int] = None,
base_url: str = "https://api.cohere.ai/v1/rerank", base_url: str = "https://api.cohere.ai/v1/rerank",
api_key: Optional[str] = None, api_key: Optional[str] = None,
**kwargs **kwargs,
) -> List[Dict[str, Any]]: ) -> List[Dict[str, Any]]:
""" """
Rerank documents using Cohere API. Rerank documents using Cohere API.
@ -254,7 +271,7 @@ async def cohere_rerank(
base_url=base_url, base_url=base_url,
api_key=api_key, api_key=api_key,
top_k=top_k, top_k=top_k,
**kwargs **kwargs,
) )
@ -266,7 +283,7 @@ async def custom_rerank(
base_url: str, base_url: str,
api_key: str, api_key: str,
top_k: Optional[int] = None, top_k: Optional[int] = None,
**kwargs **kwargs,
) -> List[Dict[str, Any]]: ) -> List[Dict[str, Any]]:
""" """
Rerank documents using a custom API endpoint. Rerank documents using a custom API endpoint.
@ -279,7 +296,7 @@ async def custom_rerank(
base_url=base_url, base_url=base_url,
api_key=api_key, api_key=api_key,
top_k=top_k, top_k=top_k,
**kwargs **kwargs,
) )
@ -297,10 +314,7 @@ if __name__ == "__main__":
query = "What is the capital of France?" query = "What is the capital of France?"
result = await jina_rerank( result = await jina_rerank(
query=query, query=query, documents=docs, top_k=2, api_key="your-api-key-here"
documents=docs,
top_k=2,
api_key="your-api-key-here"
) )
print(result) print(result)