Engram Documentation
Engram is a persistent memory layer for any AI model. It stores episodic, semantic, and procedural memories in a local SQLite database, retrieves them with sub-100 ms latency using vector search, and injects them into AI context automatically via MCP or HTTP.
What is Engram?
Modern AI assistants forget everything the moment a conversation ends. Engram solves this by maintaining a universal brain — a single, growing knowledge store shared across every tool you use.
Any AI that speaks MCP (Model Context Protocol) or HTTP can connect to Engram. Once connected, memories are automatically stored and retrieved on every interaction — no prompt engineering, no manual copy-paste, no cloud required.
Architecture
Engram runs as a local daemon (port 4901 by default). It exposes two surfaces:
- MCP server — MCP server — consumed by Claude Code, Claude Desktop, Cursor, and any MCP-compatible client
- REST API — REST API — consumed by custom apps, Ollama proxy, and direct HTTP integrations