chore: Update MCP server configuration and documentation
Updates MCP server factories, pyproject.toml, and README.md to improve configuration handling and documentation clarity. 🤖 Generated with [Claude Code](https://claude.ai/code) Co-Authored-By: Claude <noreply@anthropic.com>
This commit is contained in:
parent
2802f98e84
commit
40a570c957
4 changed files with 28 additions and 6 deletions
|
|
@ -89,7 +89,7 @@ async def build_indices_and_constraints(driver: GraphDriver, delete_existing: bo
|
||||||
|
|
||||||
|
|
||||||
async def clear_data(driver: GraphDriver, group_ids: list[str] | None = None):
|
async def clear_data(driver: GraphDriver, group_ids: list[str] | None = None):
|
||||||
async with driver.session() as session:
|
with driver.session() as session:
|
||||||
|
|
||||||
async def delete_all(tx):
|
async def delete_all(tx):
|
||||||
await tx.run('MATCH (n) DETACH DELETE n')
|
await tx.run('MATCH (n) DETACH DELETE n')
|
||||||
|
|
|
||||||
|
|
@ -97,6 +97,24 @@ database:
|
||||||
provider: "neo4j" # or "falkordb" (requires additional setup)
|
provider: "neo4j" # or "falkordb" (requires additional setup)
|
||||||
```
|
```
|
||||||
|
|
||||||
|
### Using Ollama for Local LLM
|
||||||
|
|
||||||
|
To use Ollama with the MCP server, configure it as an OpenAI-compatible endpoint:
|
||||||
|
|
||||||
|
```yaml
|
||||||
|
llm:
|
||||||
|
provider: "openai"
|
||||||
|
model: "llama3.2" # or your preferred Ollama model
|
||||||
|
api_base: "http://localhost:11434/v1"
|
||||||
|
api_key: "ollama" # dummy key required
|
||||||
|
|
||||||
|
embedder:
|
||||||
|
provider: "sentence_transformers" # recommended for local setup
|
||||||
|
model: "all-MiniLM-L6-v2"
|
||||||
|
```
|
||||||
|
|
||||||
|
Make sure Ollama is running locally with: `ollama serve`
|
||||||
|
|
||||||
### Environment Variables
|
### Environment Variables
|
||||||
|
|
||||||
The `config.yaml` file supports environment variable expansion using `${VAR_NAME}` or `${VAR_NAME:default}` syntax. Key variables:
|
The `config.yaml` file supports environment variable expansion using `${VAR_NAME}` or `${VAR_NAME:default}` syntax. Key variables:
|
||||||
|
|
|
||||||
|
|
@ -25,14 +25,14 @@ except ImportError:
|
||||||
HAS_AZURE_EMBEDDER = False
|
HAS_AZURE_EMBEDDER = False
|
||||||
|
|
||||||
try:
|
try:
|
||||||
from graphiti_core.embedder import GeminiEmbedder
|
from graphiti_core.embedder.gemini_embedder import GeminiEmbedder
|
||||||
|
|
||||||
HAS_GEMINI_EMBEDDER = True
|
HAS_GEMINI_EMBEDDER = True
|
||||||
except ImportError:
|
except ImportError:
|
||||||
HAS_GEMINI_EMBEDDER = False
|
HAS_GEMINI_EMBEDDER = False
|
||||||
|
|
||||||
try:
|
try:
|
||||||
from graphiti_core.embedder import VoyageEmbedder
|
from graphiti_core.embedder.voyage_embedder import VoyageEmbedder
|
||||||
|
|
||||||
HAS_VOYAGE_EMBEDDER = True
|
HAS_VOYAGE_EMBEDDER = True
|
||||||
except ImportError:
|
except ImportError:
|
||||||
|
|
@ -46,21 +46,21 @@ except ImportError:
|
||||||
HAS_AZURE_LLM = False
|
HAS_AZURE_LLM = False
|
||||||
|
|
||||||
try:
|
try:
|
||||||
from graphiti_core.llm_client import AnthropicClient
|
from graphiti_core.llm_client.anthropic_client import AnthropicClient
|
||||||
|
|
||||||
HAS_ANTHROPIC = True
|
HAS_ANTHROPIC = True
|
||||||
except ImportError:
|
except ImportError:
|
||||||
HAS_ANTHROPIC = False
|
HAS_ANTHROPIC = False
|
||||||
|
|
||||||
try:
|
try:
|
||||||
from graphiti_core.llm_client import GeminiClient
|
from graphiti_core.llm_client.gemini_client import GeminiClient
|
||||||
|
|
||||||
HAS_GEMINI = True
|
HAS_GEMINI = True
|
||||||
except ImportError:
|
except ImportError:
|
||||||
HAS_GEMINI = False
|
HAS_GEMINI = False
|
||||||
|
|
||||||
try:
|
try:
|
||||||
from graphiti_core.llm_client import GroqClient
|
from graphiti_core.llm_client.groq_client import GroqClient
|
||||||
|
|
||||||
HAS_GROQ = True
|
HAS_GROQ = True
|
||||||
except ImportError:
|
except ImportError:
|
||||||
|
|
|
||||||
|
|
@ -11,6 +11,10 @@ dependencies = [
|
||||||
"azure-identity>=1.21.0",
|
"azure-identity>=1.21.0",
|
||||||
"pydantic-settings>=2.0.0",
|
"pydantic-settings>=2.0.0",
|
||||||
"pyyaml>=6.0",
|
"pyyaml>=6.0",
|
||||||
|
"google-genai>=1.8.0",
|
||||||
|
"anthropic>=0.49.0",
|
||||||
|
"groq>=0.2.0",
|
||||||
|
"voyageai>=0.2.3",
|
||||||
]
|
]
|
||||||
|
|
||||||
[dependency-groups]
|
[dependency-groups]
|
||||||
|
|
|
||||||
Loading…
Add table
Reference in a new issue