ClawVault
Integrations

AI and LLMs

How ClawVault integrates with large language models for context injection, memory compression, and intelligent retrieval.

AI and LLMs

ClawVault uses LLMs in two ways: compressing raw session data into structured observations, and generating task-relevant context for prompt injection. Both are optional -- ClawVault works without any LLM access, falling back to rule-based processing.

Context Injection

The clawvault context command generates a block of relevant memories for a given task, formatted for inclusion in an LLM prompt:

clawvault context "implement OAuth for the API"

Output:

## Relevant Context (ClawVault)

### Decision: Auth Architecture (2026-02-08)
Chose OAuth 2.0 with PKCE for public clients. JWT access tokens, 15-min expiry.

### Project: API Rewrite (active)
Backend migration from Express to Fastify. Auth module is next milestone.

### Commitment: Security Audit (2026-02-28)
Pedro committed to completing auth security review before end of month.

This output is designed to be prepended to an LLM system prompt or injected as context in a conversation.

Token Budget Management

Context windows are finite. ClawVault's context command respects token budgets:

# Limit context to ~2000 tokens
clawvault context "OAuth implementation" --max-tokens 2000

# Use a specific profile for retrieval strategy
clawvault context "OAuth implementation" --profile compact

The --max-tokens flag controls how much context is returned. ClawVault prioritizes by relevance score and recency, truncating lower-priority memories first.

Context Profiles

Profiles control how context is assembled. See Context Profiles for the full reference.

ProfileBehavior
fullReturns all matching memories up to the token budget
compactSummarizes each memory to a single line
autoUses the LLM to select and rank the most relevant memories
recentPrioritizes chronologically recent memories

The auto profile makes an LLM call to intelligently select which memories matter for the given task. This produces the best results but adds latency and cost:

clawvault context "deploy to production" --profile auto

Observational Memory Compression

The observe --compress command uses an LLM to read raw session transcripts and extract structured observations:

clawvault observe --compress

This produces categorized observations with scored importance using the [type|c=confidence|i=importance] format:

  • Structural (importance >= 0.8) -- decisions made, errors encountered, deadlines set
  • Potential (importance 0.4-0.79) -- preferences expressed, architecture discussions, people interactions
  • Contextual (importance < 0.4) -- routine updates, successful deployments, progress notes

Which Models Work

Observation compression works with any model that handles long context well. Recommended:

ProviderModelNotes
AnthropicClaude 3.5 Sonnet or laterBest quality, handles long transcripts well
OpenAIGPT-4oGood balance of speed and quality
GoogleGemini 1.5 ProLarge context window, good for very long sessions

Set the API key for your preferred provider:

export ANTHROPIC_API_KEY="sk-ant-..."
# or
export OPENAI_API_KEY="sk-..."
# or
export GEMINI_API_KEY="..."

ClawVault detects which keys are available and uses the first one found (in the order above).

Without any API key, observe falls back to rule-based extraction. This catches obvious patterns (errors, TODOs, decisions marked with keywords) but misses nuanced context that LLM compression would capture.

Supported Providers

ClawVault supports three LLM providers:

Anthropic

export ANTHROPIC_API_KEY="sk-ant-..."
export CLAWVAULT_MODEL="claude-sonnet-4-20250514"  # optional, uses default if unset

OpenAI

export OPENAI_API_KEY="sk-..."
export CLAWVAULT_MODEL="gpt-4o"  # optional

Google Gemini

export GEMINI_API_KEY="..."
export CLAWVAULT_MODEL="gemini-1.5-pro"  # optional

The CLAWVAULT_MODEL environment variable overrides the default model for the detected provider. If unset, ClawVault uses a sensible default for each provider.

Running Without an LLM

ClawVault is fully functional without LLM access. The following features degrade gracefully:

FeatureWith LLMWithout LLM
observeAI-compressed observationsRule-based extraction
context --profile autoLLM-ranked relevanceFalls back to full profile
wakeAI-generated session summaryRaw handoff + recent memories
SearchUnaffectedUnaffected
StorageUnaffectedUnaffected
GraphUnaffectedUnaffected

This means ClawVault works in air-gapped environments, on machines without internet access, or when you simply prefer not to send vault contents to an API.

On this page