auto-claude-memory
Original:🇺🇸 English
Translated
Auto-Claude Graphiti memory system configuration and usage. Use when setting up memory persistence, configuring LLM/embedding providers, querying knowledge graph, or optimizing memory performance.
2installs
Sourceadaptationio/skrillz
Added on
NPX Install
npx skill4agent add adaptationio/skrillz auto-claude-memoryTags
Translated version includes tags in frontmatterSKILL.md Content
View Translation Comparison →Auto-Claude Memory System
Graphiti-based persistent memory for cross-session context retention.
Overview
Auto-Claude uses Graphiti with embedded LadybugDB for memory:
- No Docker required - Embedded graph database
- Multi-provider support - OpenAI, Anthropic, Ollama, Google AI, Azure
- Semantic search - Find relevant context across sessions
- Knowledge graph - Entity relationships and facts
Architecture
Agent Session
│
▼
Memory Manager
│
├──▶ Add Episode (new learnings)
├──▶ Search Nodes (find entities)
├──▶ Search Facts (find relationships)
└──▶ Get Context (relevant memories)
│
▼
Graphiti (Knowledge Graph)
│
▼
LadybugDB (Embedded Storage)Configuration
Enable Memory System
In :
apps/backend/.envbash
# Enable Graphiti memory (default: true)
GRAPHITI_ENABLED=trueProvider Selection
Choose LLM and embedding providers:
bash
# LLM provider: openai | anthropic | azure_openai | ollama | google | openrouter
GRAPHITI_LLM_PROVIDER=openai
# Embedder provider: openai | voyage | azure_openai | ollama | google | openrouter
GRAPHITI_EMBEDDER_PROVIDER=openaiProvider Configurations
OpenAI (Simplest)
bash
GRAPHITI_ENABLED=true
GRAPHITI_LLM_PROVIDER=openai
GRAPHITI_EMBEDDER_PROVIDER=openai
OPENAI_API_KEY=sk-xxxxxxxxxxxxxxxx
OPENAI_MODEL=gpt-4o-mini
OPENAI_EMBEDDING_MODEL=text-embedding-3-smallAnthropic + Voyage (High Quality)
bash
GRAPHITI_ENABLED=true
GRAPHITI_LLM_PROVIDER=anthropic
GRAPHITI_EMBEDDER_PROVIDER=voyage
ANTHROPIC_API_KEY=sk-ant-xxxxxxxx
GRAPHITI_ANTHROPIC_MODEL=claude-sonnet-4-5-latest
VOYAGE_API_KEY=pa-xxxxxxxx
VOYAGE_EMBEDDING_MODEL=voyage-3Ollama (Fully Offline)
bash
GRAPHITI_ENABLED=true
GRAPHITI_LLM_PROVIDER=ollama
GRAPHITI_EMBEDDER_PROVIDER=ollama
OLLAMA_BASE_URL=http://localhost:11434
OLLAMA_LLM_MODEL=deepseek-r1:7b
OLLAMA_EMBEDDING_MODEL=nomic-embed-text
OLLAMA_EMBEDDING_DIM=768Prerequisites:
bash
# Install Ollama
curl -fsSL https://ollama.ai/install.sh | sh
# Pull models
ollama pull deepseek-r1:7b
ollama pull nomic-embed-textGoogle AI (Gemini)
bash
GRAPHITI_ENABLED=true
GRAPHITI_LLM_PROVIDER=google
GRAPHITI_EMBEDDER_PROVIDER=google
GOOGLE_API_KEY=AIzaSyxxxxxxxx
GOOGLE_LLM_MODEL=gemini-2.0-flash
GOOGLE_EMBEDDING_MODEL=text-embedding-004Azure OpenAI (Enterprise)
bash
GRAPHITI_ENABLED=true
GRAPHITI_LLM_PROVIDER=azure_openai
GRAPHITI_EMBEDDER_PROVIDER=azure_openai
AZURE_OPENAI_API_KEY=xxxxxxxx
AZURE_OPENAI_BASE_URL=https://your-resource.openai.azure.com/...
AZURE_OPENAI_LLM_DEPLOYMENT=gpt-4
AZURE_OPENAI_EMBEDDING_DEPLOYMENT=text-embedding-3-smallOpenRouter (Multi-Provider)
bash
GRAPHITI_ENABLED=true
GRAPHITI_LLM_PROVIDER=openrouter
GRAPHITI_EMBEDDER_PROVIDER=openrouter
OPENROUTER_API_KEY=sk-or-xxxxxxxx
OPENROUTER_LLM_MODEL=anthropic/claude-3.5-sonnet
OPENROUTER_EMBEDDING_MODEL=openai/text-embedding-3-smallDatabase Settings
bash
# Database name (default: auto_claude_memory)
GRAPHITI_DATABASE=auto_claude_memory
# Storage path (default: ~/.auto-claude/memories)
GRAPHITI_DB_PATH=~/.auto-claude/memoriesMemory Operations
How Memory Works
-
During Build
- Agent discovers patterns, gotchas, solutions
- Memory Manager extracts insights
- Insights stored as episodes in knowledge graph
-
New Session
- Agent queries for relevant context
- Memory returns related insights
- Agent builds on previous learnings
MCP Tools
When is set, agents can use:
GRAPHITI_MCP_URL| Tool | Purpose |
|---|---|
| Search entity summaries |
| Search relationships between entities |
| Add data to knowledge graph |
| Retrieve recent episodes |
| Get specific entity/relationship |
Python API
python
from integrations.graphiti.memory import get_graphiti_memory
# Get memory instance
memory = get_graphiti_memory(spec_dir, project_dir)
# Get context for session
context = memory.get_context_for_session("Implementing feature X")
# Add insight from session
memory.add_session_insight("Pattern: use React hooks for state")
# Search for relevant memories
results = memory.search("authentication patterns")Memory Storage
Location
~/.auto-claude/memories/
├── auto_claude_memory/ # Main database
│ ├── nodes/ # Entity nodes
│ ├── edges/ # Relationships
│ └── episodes/ # Session insights
└── embeddings/ # Vector embeddingsPer-Spec Memory
.auto-claude/specs/001-feature/
└── graphiti/ # Spec-specific memory
├── insights.json # Extracted insights
└── context.json # Session contextQuerying Memory
Command Line
bash
cd apps/backend
# Query memory
python query_memory.py --search "authentication"
# List recent episodes
python query_memory.py --recent 10
# Get entity details
python query_memory.py --entity "UserService"Memory in Action
Example session:
Session 1:
Agent: "Implemented OAuth login, discovered need to handle token refresh"
Memory: Stores insight about token refresh pattern
Session 2:
Agent: "Implementing user profile..."
Memory: "Previously learned about token refresh in OAuth implementation"
Agent: Uses learned pattern for profile API callsBest Practices
Effective Memory Use
-
Let agents learn naturally
- Don't force memory storage
- Agents automatically extract insights
-
Use semantic search
- Query with natural language
- Memory finds related concepts
-
Clean up periodically
- Remove outdated insights
- Update incorrect information
Provider Selection
| Use Case | Recommended |
|---|---|
| Production | OpenAI or Anthropic+Voyage |
| Development | Ollama (free, offline) |
| Enterprise | Azure OpenAI |
| Budget | OpenRouter or Google AI |
Performance Tips
-
Embedding model selection
- : Fast, good quality
text-embedding-3-small - : Better quality, slower
text-embedding-3-large
-
LLM model selection
- : Fast, cost-effective
gpt-4o-mini - : High quality reasoning
claude-sonnet
-
Ollama optimizationbash
# Use smaller models for speed OLLAMA_LLM_MODEL=llama3.2:3b OLLAMA_EMBEDDING_MODEL=all-minilm OLLAMA_EMBEDDING_DIM=384
Troubleshooting
Memory Not Working
bash
# Check if enabled
grep GRAPHITI apps/backend/.env
# Verify provider credentials
python -c "from integrations.graphiti.memory import get_graphiti_memory; print('OK')"Provider Errors
bash
# OpenAI
curl -H "Authorization: Bearer $OPENAI_API_KEY" https://api.openai.com/v1/models
# Ollama
curl http://localhost:11434/api/tags
# Check logs
DEBUG=true python query_memory.py --search "test"Database Corruption
bash
# Backup and reset
mv ~/.auto-claude/memories ~/.auto-claude/memories.backup
python query_memory.py --search "test" # Creates fresh DBEmbedding Dimension Mismatch
If changing embedding models:
bash
# Clear existing embeddings
rm -rf ~/.auto-claude/memories/embeddings
# Restart to re-embed
python run.py --spec 001Advanced Usage
Custom Memory Integration
python
from integrations.graphiti.queries_pkg.graphiti import GraphitiMemory
# Create custom memory instance
memory = GraphitiMemory(
database="custom_db",
db_path="/path/to/storage",
llm_provider="anthropic",
embedder_provider="voyage"
)
# Custom operations
memory.add_entity("UserService", {"type": "service", "purpose": "auth"})
memory.add_relationship("UserService", "uses", "Database")Memory MCP Server
Run standalone memory server:
bash
# Start Graphiti MCP server
GRAPHITI_MCP_URL=http://localhost:8000/mcp/ python -m integrations.graphiti.serverRelated Skills
- auto-claude-setup: Initial configuration
- auto-claude-optimization: Performance tuning
- auto-claude-troubleshooting: Debugging