fix: handle provider prefix in LiteLLMEmbeddingEngine tokenizer loading
Signed-off-by: Faizan Shaikh <faizansk9292@gmail.com>
This commit is contained in:
parent
1cf0a202ce
commit
f637f80d7a
1 changed files with 2 additions and 1 deletions
|
|
@ -219,7 +219,7 @@ class LiteLLMEmbeddingEngine(EmbeddingEngine):
|
||||||
else:
|
else:
|
||||||
try:
|
try:
|
||||||
tokenizer = HuggingFaceTokenizer(
|
tokenizer = HuggingFaceTokenizer(
|
||||||
model=self.model.replace("hosted_vllm/", ""),
|
model=self.model.replace("hosted_vllm/", "").replace("openai/", ""),
|
||||||
max_completion_tokens=self.max_completion_tokens,
|
max_completion_tokens=self.max_completion_tokens,
|
||||||
)
|
)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
|
|
@ -231,3 +231,4 @@ class LiteLLMEmbeddingEngine(EmbeddingEngine):
|
||||||
|
|
||||||
logger.debug(f"Tokenizer loaded for model: {self.model}")
|
logger.debug(f"Tokenizer loaded for model: {self.model}")
|
||||||
return tokenizer
|
return tokenizer
|
||||||
|
|
||||||
|
|
|
||||||
Loading…
Add table
Reference in a new issue