diff --git a/lightrag/operate.py b/lightrag/operate.py index b0736402..3a7502c1 100644 --- a/lightrag/operate.py +++ b/lightrag/operate.py @@ -72,8 +72,8 @@ def chunking_by_token_size( content: str, split_by_character: str | None = None, split_by_character_only: bool = False, - chunk_overlap_token_size: int = 128, - chunk_token_size: int = 1024, + chunk_overlap_token_size: int = 100, + chunk_token_size: int = 1200, ) -> list[dict[str, Any]]: tokens = tokenizer.encode(content) results: list[dict[str, Any]] = []