reformat
This commit is contained in:
parent
e5341c5f49
commit
8965e31a58
1 changed files with 4 additions and 5 deletions
|
|
@ -188,17 +188,16 @@ def get_llm_client(raise_api_key_error: bool = True):
|
||||||
instructor_mode=llm_config.llm_instructor_mode.lower(),
|
instructor_mode=llm_config.llm_instructor_mode.lower(),
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
elif provider == LLMProvider.LLAMA_CPP:
|
elif provider == LLMProvider.LLAMA_CPP:
|
||||||
from cognee.infrastructure.llm.structured_output_framework.litellm_instructor.llm.llama_cpp.adapter import (
|
from cognee.infrastructure.llm.structured_output_framework.litellm_instructor.llm.llama_cpp.adapter import (
|
||||||
LlamaCppAPIAdapter,
|
LlamaCppAPIAdapter,
|
||||||
)
|
)
|
||||||
|
|
||||||
# Get optional local mode parameters (will be None if not set)
|
# Get optional local mode parameters (will be None if not set)
|
||||||
model_path = getattr(llm_config, 'llama_cpp_model_path', None)
|
model_path = getattr(llm_config, "llama_cpp_model_path", None)
|
||||||
n_ctx = getattr(llm_config, 'llama_cpp_n_ctx', 2048)
|
n_ctx = getattr(llm_config, "llama_cpp_n_ctx", 2048)
|
||||||
n_gpu_layers = getattr(llm_config, 'llama_cpp_n_gpu_layers', 0)
|
n_gpu_layers = getattr(llm_config, "llama_cpp_n_gpu_layers", 0)
|
||||||
chat_format = getattr(llm_config, 'llama_cpp_chat_format', 'chatml')
|
chat_format = getattr(llm_config, "llama_cpp_chat_format", "chatml")
|
||||||
|
|
||||||
return LlamaCppAPIAdapter(
|
return LlamaCppAPIAdapter(
|
||||||
model=llm_config.llm_model,
|
model=llm_config.llm_model,
|
||||||
|
|
|
||||||
Loading…
Add table
Reference in a new issue