Reduce default embedding batch size from 32 to 10
This commit is contained in:
parent
186c6411fb
commit
9b7b2a9b0f
2 changed files with 2 additions and 2 deletions
|
|
@ -96,7 +96,7 @@ EMBEDDING_BINDING_API_KEY=your_api_key
|
|||
# If the embedding service is deployed within the same Docker stack, use host.docker.internal instead of localhost
|
||||
EMBEDDING_BINDING_HOST=http://localhost:11434
|
||||
### Num of chunks send to Embedding in single request
|
||||
# EMBEDDING_BATCH_NUM=32
|
||||
# EMBEDDING_BATCH_NUM=10
|
||||
### Max concurrency requests for Embedding
|
||||
# EMBEDDING_FUNC_MAX_ASYNC=16
|
||||
### Maximum tokens sent to Embedding for each chunk (no longer in use?)
|
||||
|
|
|
|||
|
|
@ -201,7 +201,7 @@ class LightRAG:
|
|||
embedding_func: EmbeddingFunc | None = field(default=None)
|
||||
"""Function for computing text embeddings. Must be set before use."""
|
||||
|
||||
embedding_batch_num: int = field(default=int(os.getenv("EMBEDDING_BATCH_NUM", 32)))
|
||||
embedding_batch_num: int = field(default=int(os.getenv("EMBEDDING_BATCH_NUM", 10)))
|
||||
"""Batch size for embedding computations."""
|
||||
|
||||
embedding_func_max_async: int = field(
|
||||
|
|
|
|||
Loading…
Add table
Reference in a new issue