From 19419281c32377f9125add2d3d0fc65ab345cb40 Mon Sep 17 00:00:00 2001 From: Stephen Hu Date: Wed, 9 Jul 2025 12:17:26 +0800 Subject: [PATCH] Fix: Change Ollama Embedding Keep Alive (#8734) ### What problem does this PR solve? https://github.com/infiniflow/ragflow/issues/8733 ### Type of change - [x] Bug Fix (non-breaking change which fixes an issue) --- rag/llm/embedding_model.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/rag/llm/embedding_model.py b/rag/llm/embedding_model.py index 1b0b927b0..6530e9b67 100644 --- a/rag/llm/embedding_model.py +++ b/rag/llm/embedding_model.py @@ -286,7 +286,7 @@ class OllamaEmbed(Base): # remove special tokens if they exist for token in OllamaEmbed._special_tokens: txt = txt.replace(token, "") - res = self.client.embeddings(prompt=txt, model=self.model_name, options={"use_mmap": True}) + res = self.client.embeddings(prompt=txt, model=self.model_name, options={"use_mmap": True}, keep_alive=-1) try: arr.append(res["embedding"]) except Exception as _e: @@ -298,7 +298,7 @@ class OllamaEmbed(Base): # remove special tokens if they exist for token in OllamaEmbed._special_tokens: text = text.replace(token, "") - res = self.client.embeddings(prompt=text, model=self.model_name, options={"use_mmap": True}) + res = self.client.embeddings(prompt=text, model=self.model_name, options={"use_mmap": True}, keep_alive=-1) try: return np.array(res["embedding"]), 128 except Exception as _e: