Merge branch 'main' into fix-issue-1746
This commit is contained in:
commit
b6ab69e25d
3 changed files with 13 additions and 4 deletions
|
|
@ -30,7 +30,7 @@
|
|||
<a href="https://github.com/HKUDS/LightRAG/issues/285"><img src="https://img.shields.io/badge/💬微信群-交流-07c160?style=for-the-badge&logo=wechat&logoColor=white&labelColor=1a1a2e"></a>
|
||||
</p>
|
||||
<p>
|
||||
<a href="README_zh.md"><img src="https://img.shields.io/badge/🇨🇳中文版-1a1a2e?style=for-the-badge"></a>
|
||||
<a href="README-zh.md"><img src="https://img.shields.io/badge/🇨🇳中文版-1a1a2e?style=for-the-badge"></a>
|
||||
<a href="README.md"><img src="https://img.shields.io/badge/🇺🇸English-1a1a2e?style=for-the-badge"></a>
|
||||
</p>
|
||||
</div>
|
||||
|
|
|
|||
|
|
@ -26,6 +26,7 @@ from .utils import (
|
|||
get_conversation_turns,
|
||||
use_llm_func_with_cache,
|
||||
update_chunk_cache_list,
|
||||
remove_think_tags,
|
||||
)
|
||||
from .base import (
|
||||
BaseGraphStorage,
|
||||
|
|
@ -1703,7 +1704,8 @@ async def extract_keywords_only(
|
|||
result = await use_model_func(kw_prompt, keyword_extraction=True)
|
||||
|
||||
# 6. Parse out JSON from the LLM response
|
||||
match = re.search(r"\{.*\}", result, re.DOTALL)
|
||||
result = remove_think_tags(result)
|
||||
match = re.search(r"\{.*?\}", result, re.DOTALL)
|
||||
if not match:
|
||||
logger.error("No JSON-like structure found in the LLM respond.")
|
||||
return [], []
|
||||
|
|
|
|||
|
|
@ -1465,6 +1465,11 @@ async def update_chunk_cache_list(
|
|||
)
|
||||
|
||||
|
||||
def remove_think_tags(text: str) -> str:
|
||||
"""Remove <think> tags from the text"""
|
||||
return re.sub(r"^(<think>.*?</think>|<think>)", "", text, flags=re.DOTALL).strip()
|
||||
|
||||
|
||||
async def use_llm_func_with_cache(
|
||||
input_text: str,
|
||||
use_llm_func: callable,
|
||||
|
|
@ -1531,6 +1536,7 @@ async def use_llm_func_with_cache(
|
|||
kwargs["max_tokens"] = max_tokens
|
||||
|
||||
res: str = await use_llm_func(input_text, **kwargs)
|
||||
res = remove_think_tags(res)
|
||||
|
||||
if llm_response_cache.global_config.get("enable_llm_cache_for_entity_extract"):
|
||||
await save_to_cache(
|
||||
|
|
@ -1557,8 +1563,9 @@ async def use_llm_func_with_cache(
|
|||
if max_tokens is not None:
|
||||
kwargs["max_tokens"] = max_tokens
|
||||
|
||||
logger.info(f"Call LLM function with query text lenght: {len(input_text)}")
|
||||
return await use_llm_func(input_text, **kwargs)
|
||||
logger.info(f"Call LLM function with query text length: {len(input_text)}")
|
||||
res = await use_llm_func(input_text, **kwargs)
|
||||
return remove_think_tags(res)
|
||||
|
||||
|
||||
def get_content_summary(content: str, max_length: int = 250) -> str:
|
||||
|
|
|
|||
Loading…
Add table
Reference in a new issue