add async for related_questions
This commit is contained in:
parent
d5139b10ec
commit
0d7537b852
3 changed files with 11 additions and 6 deletions
|
|
@ -402,7 +402,7 @@ async def related_questions():
|
||||||
if "parameter" in gen_conf:
|
if "parameter" in gen_conf:
|
||||||
del gen_conf["parameter"]
|
del gen_conf["parameter"]
|
||||||
prompt = load_prompt("related_question")
|
prompt = load_prompt("related_question")
|
||||||
ans = chat_mdl.chat(
|
ans = await chat_mdl.async_chat(
|
||||||
prompt,
|
prompt,
|
||||||
[
|
[
|
||||||
{
|
{
|
||||||
|
|
|
||||||
|
|
@ -788,7 +788,7 @@ Reason:
|
||||||
- At the same time, related terms can also help search engines better understand user needs and return more accurate search results.
|
- At the same time, related terms can also help search engines better understand user needs and return more accurate search results.
|
||||||
|
|
||||||
"""
|
"""
|
||||||
ans = chat_mdl.chat(
|
ans = await chat_mdl.async_chat(
|
||||||
prompt,
|
prompt,
|
||||||
[
|
[
|
||||||
{
|
{
|
||||||
|
|
@ -1070,7 +1070,7 @@ async def related_questions_embedded():
|
||||||
|
|
||||||
gen_conf = search_config.get("llm_setting", {"temperature": 0.9})
|
gen_conf = search_config.get("llm_setting", {"temperature": 0.9})
|
||||||
prompt = load_prompt("related_question")
|
prompt = load_prompt("related_question")
|
||||||
ans = chat_mdl.chat(
|
ans = await chat_mdl.async_chat(
|
||||||
prompt,
|
prompt,
|
||||||
[
|
[
|
||||||
{
|
{
|
||||||
|
|
|
||||||
|
|
@ -1765,12 +1765,17 @@ class LiteLLMBase(ABC):
|
||||||
|
|
||||||
yield ans, tol
|
yield ans, tol
|
||||||
|
|
||||||
async def async_chat(self, history, gen_conf, **kwargs):
|
async def async_chat(self, system, history, gen_conf, **kwargs):
|
||||||
logging.info("[HISTORY]" + json.dumps(history, ensure_ascii=False, indent=2))
|
hist = list(history) if history else []
|
||||||
|
if system:
|
||||||
|
if not hist or hist[0].get("role") != "system":
|
||||||
|
hist.insert(0, {"role": "system", "content": system})
|
||||||
|
|
||||||
|
logging.info("[HISTORY]" + json.dumps(hist, ensure_ascii=False, indent=2))
|
||||||
if self.model_name.lower().find("qwen3") >= 0:
|
if self.model_name.lower().find("qwen3") >= 0:
|
||||||
kwargs["extra_body"] = {"enable_thinking": False}
|
kwargs["extra_body"] = {"enable_thinking": False}
|
||||||
|
|
||||||
completion_args = self._construct_completion_args(history=history, stream=False, tools=False, **gen_conf)
|
completion_args = self._construct_completion_args(history=hist, stream=False, tools=False, **gen_conf)
|
||||||
|
|
||||||
for attempt in range(self.max_retries + 1):
|
for attempt in range(self.max_retries + 1):
|
||||||
try:
|
try:
|
||||||
|
|
|
||||||
Loading…
Add table
Reference in a new issue