┌─ palinode ─┐
│ ░░░░░░░░░░ │
│ ▓▓▓▓▓▓▓▓▓▓ │
│ ██████████ │
└────────────┘
The memory substrate for AI agents and developer tools. Git-versioned, file-native, MCP-first.
Your agent's memory is a folder of markdown files. Palinode indexes them with hybrid search, compacts them with an LLM, and serves them through MCP — so the same memory works in Claude Code, Cursor, Windsurf, Zed, VS Code (Continue/Cline), and any other MCP-compatible editor. Enterprises can govern AI memory the same way they govern code. If every service crashes, cat still works.
A palinode is a poem that retracts what was said before and says it better. That's what memory compaction does.
Most agent memory is a black box. You can't read it, you can't diff it, you can't grep it when the vector DB is down. Palinode bets on plain files as the source of truth and builds everything else as a derived index.
Files (markdown + YAML frontmatter)
↓ watched
Index (SQLite-vec vectors + FTS5 keywords, single .db file)
↓ queried by
Interfaces (MCP server, REST API, CLI, OpenClaw plugin)
↓ compacted by
LLM (proposes ops → deterministic executor applies them → git commits)
That's the whole architecture. One directory of .md files, one SQLite database, one API server. No Postgres, no Redis, no cloud dependency.
Palinode doesn't care how you talk to it. The same 17 tools work everywhere:
| Interface | Transport | Best For |
|---|---|---|
| MCP Server | Streamable HTTP or stdio | Claude Code, Claude Desktop, Cursor, Windsurf, Zed, VS Code (Continue/Cline) |
| REST API | HTTP on :6340 | Scripts, webhooks, custom integrations |
| CLI | Wraps REST API | Cron jobs, SSH, shell scripts (8x fewer tokens than MCP) |
| Plugin | OpenClaw lifecycle hooks | Agent frameworks with inject/extract patterns |
Set up once on a server. Connect from any machine, any IDE, any agent framework. The MCP server is a pure HTTP client — it holds no state, no database connection, no embedder. Point it at the API and go.
{
"mcpServers": {
"palinode": { "url": "http://your-server:6341/mcp/" }
}
}That's the entire client config. Works with Claude Code, Claude Desktop, Cursor, Windsurf, Zed, and VS Code (Continue/Cline). See docs/MCP-SETUP.md for editor-specific paths.
Store — Typed markdown files (people, projects, decisions, insights) with YAML frontmatter. Git-versioned. Human-readable. Editable in Obsidian, VS Code, vim, or anything.
Index — A file watcher embeds with BGE-M3 and indexes with FTS5 as you save. Content-hash dedup skips re-embedding unchanged files (~90% savings). Single SQLite file, zero external services.
Search — Hybrid BM25 + vector search merged with Reciprocal Rank Fusion. Keyword precision when you need exact terms, semantic recall when you don't. Optional associative entity graph and prospective triggers.
Compact — Weekly consolidation where an LLM proposes structured operations (KEEP / UPDATE / MERGE / SUPERSEDE / ARCHIVE) and a deterministic executor applies them. The LLM never touches your files directly. Every compaction is a git commit you can review, blame, or revert.
Audit — git blame any fact. git diff any change. rollback any mistake. These aren't just git-compatible files — palinode_diff, palinode_blame, and palinode_rollback are first-class tools your agent can call.
Already have Palinode installed and palinode-api running? Drop it into any
project in one command:
cd your-project
palinode initThat scaffolds:
.claude/CLAUDE.md— memory instructions for the agent (appended if one already exists).claude/settings.json— aSessionEndhook that auto-captures on/clear, logout, and normal exit.claude/hooks/palinode-session-end.sh— the hook script itself.mcp.json— points Claude Code at thepalinodeMCP server
Open the project in Claude Code and your agent will search prior context on
startup, save decisions as you work, and snapshot the session on /clear. No
server restarts, no settings menus, no copy-paste.
Re-run with --dry-run to preview, --force to overwrite, or --no-mcp
/ --no-hook to scope what gets installed.
# Install
git clone https://github.com/phasespace-labs/palinode && cd palinode
python3 -m venv venv && source venv/bin/activate
pip install -e .
# Create your memory directory
mkdir -p ~/.palinode/{people,projects,decisions,insights,daily}
cd ~/.palinode && git init
cp /path/to/palinode/palinode.config.yaml.example palinode.config.yaml # adjust path
# Start services
PALINODE_DIR=~/.palinode palinode-api # REST API on :6340
PALINODE_DIR=~/.palinode palinode-watcher # auto-indexes on file save
PALINODE_DIR=~/.palinode palinode-mcp-sse # MCP server on :6341 (optional)
# Verify
curl http://localhost:6340/statusYour memory directory is private. It contains personal data. Never make it public. The code repo contains zero memory files.
For a pre-populated demo, copy
examples/sample-memory/to~/.palinode/.
# During a session — save a decision
palinode save --type Decision "Chose SQLite over Postgres for the cache layer. \
Reason: no ops burden, single-file deployment, good enough for our scale."
# Next week — search for it
palinode search "database decision for cache"# Agent calls at end of coding session
palinode session-end \
--summary "Migrated auth from JWT to session tokens" \
--decisions "Session tokens stored server-side, 24h expiry" \
--blockers "Need to update mobile client auth flow"# Trace a fact back to when it was recorded
palinode blame decisions/auth-migration.md
# See what changed across all memory in the last week
palinode diff --days 717 tools available through every interface:
| Tool | What It Does |
|---|---|
search |
Hybrid BM25 + vector search with category filter |
save |
Store a typed memory (person, decision, insight, project) |
list |
Browse memory files by type, filter by core status |
read |
Read the full content of a memory file |
ingest |
Fetch a URL and save as research |
status |
Health check — file counts, index stats, service status |
entities |
Entity graph — cross-references between memories |
consolidate |
Preview or run LLM-powered compaction |
diff |
What changed in the last N days |
blame |
Trace a fact back to the commit that recorded it |
history |
Git history for a file with diff stats and rename tracking |
rollback |
Revert a file to a previous commit (safe, creates new commit) |
push |
Sync memory to a remote git repo |
trigger |
Prospective recall — auto-inject when a topic comes up |
lint |
Health scan — orphans, stale files, missing fields |
session_end |
Capture summary, decisions, and blockers at end of session |
prompt |
List, show, or activate versioned LLM prompts |
Every tool is accessible as palinode_<name> via MCP, palinode <name> via CLI, or POST/GET /<name> via the REST API.
| Layer | Choice | Why |
|---|---|---|
| Source of truth | Markdown + YAML frontmatter | Human-readable, git-versioned, portable |
| Vector index | SQLite-vec (embedded) | No server, single file, zero config |
| Keyword index | SQLite FTS5 (embedded) | BM25 for exact terms, zero dependencies |
| Embeddings | BGE-M3 via Ollama | Local, private, no API key needed |
| API | FastAPI | Lightweight, async, one process |
| MCP | Python MCP SDK (Streamable HTTP) | Works with every IDE over the network |
| CLI | Click (wraps REST API) | Shell-native, TTY-aware output |
| Behavior | PROGRAM.md |
What to remember, how to extract, how to compact — edit one file to change all behavior |
---
id: project-palinode
category: project
name: Palinode
core: true
status: active
entities: [person/paul]
last_updated: 2026-04-05T00:00:00Z
summary: "Persistent memory for AI agents."
canonical_question: "What is Palinode and what does it do?"
---
# Palinode
Your content here. As detailed or brief as you want.
Files marked `core: true` are always in context.
Everything else is retrieved on demand via hybrid search.
The `canonical_question` field anchors the file to the question it answers, improving search relevance.Open your memory directory as an Obsidian vault for visual browsing. See docs/OBSIDIAN-SETUP.md.
All behavior is in palinode.config.yaml:
memory_dir: "~/.palinode"
ollama_url: "http://localhost:11434"
embedding_model: "bge-m3"
recall:
search:
top_k: 5
threshold: 0.4
core:
max_chars_per_file: 3000
search:
hybrid_enabled: true
hybrid_weight: 0.5 # 0.0 = vector only, 1.0 = BM25 only
consolidation:
llm_model: "llama3.1:8b" # any chat model that outputs JSON
llm_url: "http://localhost:11434"
llm_fallbacks: # tried in order if primary fails
- model: "qwen2.5:14b-instruct"
url: "http://localhost:11434"All models are swappable. Any Ollama embedding model, any OpenAI-compatible chat endpoint. See palinode.config.yaml.example for the full reference.
- Python 3.11+
- Ollama with
bge-m3(ollama pull bge-m3) - Git
Optional: a chat model for consolidation (any 7B+ works), OpenClaw for agent plugin hooks.
| Method | Path | Description |
|---|---|---|
GET |
/status |
Health check + stats |
POST |
/search |
Hybrid search with filters |
POST |
/search-associative |
Entity graph traversal |
POST |
/save |
Create a typed memory file |
POST |
/ingest-url |
Fetch URL, save as research |
GET/POST |
/triggers |
Prospective recall triggers |
POST |
/consolidate |
Run or preview compaction |
GET |
/list |
Browse files by type |
GET |
/read?file_path=... |
Read a memory file |
GET |
/history/{file_path} |
Git log for a file |
GET |
/diff |
Recent changes |
GET |
/blame/{file_path} |
Git blame |
POST |
/rollback |
Revert a file |
POST |
/push |
Push to git remote |
POST |
/reindex |
Rebuild indices |
POST |
/session-end |
Capture session summary |
POST |
/lint |
Health scan |
-
Files are truth. Not databases, not vector stores. Markdown files that humans can read, edit, and version with git.
-
Typed, not flat. People, projects, decisions, insights — each has structure. This enables reliable retrieval and consolidation.
-
Consolidation, not accumulation. 100 sessions should produce 20 well-maintained files, not 100 unread dumps.
-
Invisible when working. The human talks to their agent. Palinode works behind the scenes.
-
Graceful degradation. Vector index down? Read files directly. Embedding service down? Grep. Machine off? It's a git repo, clone it anywhere.
-
Zero taxonomy burden. The system classifies. The human reviews. If the human has to maintain a taxonomy, the system dies.
- Your data, your files — No accounts, no cloud dependency, no vendor lock-in. Your memory is markdown files in a directory you control. Export is
cp. Backup isgit push. Whatever happens to any tool in this ecosystem, your data is plain text on your filesystem. - Cross-IDE memory — Your memory lives in one place. Connect from Claude Code, Cursor, Windsurf, Zed, or any MCP-compatible editor. Switch IDEs without losing context.
- Git operations as agent tools —
diff,blame,rollback,pushexposed via MCP. No other system makes git ops callable by the agent. - Operation-based compaction — KEEP/UPDATE/MERGE/SUPERSEDE/ARCHIVE DSL. LLM proposes, deterministic executor disposes. Every compaction is a reviewable git commit.
- Per-fact addressability —
<!-- fact:slug -->IDs inline in markdown, invisible in rendering, preserved by git, targetable by compaction. - 4-phase injection — Core (always) + Topic (per-turn search) + Associative (entity graph) + Triggered (prospective recall).
- Multi-transport MCP — stdio for local, Streamable HTTP for remote. One server, any IDE on any machine.
- If everything crashes,
catstill works.
Palinode builds on ideas from Karpathy's LLM Knowledge Bases, Letta (tiered memory), and LangMem (typed schemas + background consolidation). See docs/ACKNOWLEDGMENTS.md for the full list.
See also the epistemic integrity discussion in the Karpathy gist thread — particularly the problem of LLM wikis that "synthesise without citing, drift from sources without knowing it, and present false certainty where disagreement exists." Git-based provenance is Palinode's answer to that problem.
If you know of prior art we missed, please open an issue.
MIT — Privacy Policy
Built by Paul Kyle with help from AI agents who use Palinode to remember building Palinode.