Memory that thinks like you do.
A plugin-based memory SDK for LLM applications. Remote storage, local runtime, zero latency.
Everything you need to give your AI application persistent, intelligent memory.
Four-layer architecture mirrors how humans remember: working, short-term, long-term, and archive. Memories decay, consolidate, and strengthen with use.
All memory intelligence is pluggable. Use official plugins for extraction, retrieval, decay, consolidation, and conflict detection — or build your own.
Git-like architecture: pull memories once, run everything locally. No network round-trip on retrieval. Works offline, syncs when connected.
Switch devices, switch projects. Your memories follow you. One API key, one init() call — full context restored.
Use your own API keys for Anthropic, OpenAI, or any OpenAI-compatible provider. Server has zero LLM cost.
Built-in MCP server adapter. Connect to Claude Desktop, Cursor, or any MCP-compatible client with zero code.
SDK runs locally for zero latency. Server handles storage and sync. Like Git for memory.
No vector database. No graph database. No server-side LLM cost.
| Mem0 | Zep | Letta | Inkess Memory | |
|---|---|---|---|---|
| Storage | Vector DB | Graph + Vector | PostgreSQL | File Storage |
| LLM Cost | Server-side | Server-side | Server-side | Client-side (yours) |
| Latency | Network | Network | Low | Zero |
| Offline | ✗ | ✗ | ✓ | ✓ |
| Memory Model | Flat vector | Temporal graph | OS memory blocks | Human-like layers |
| Plugin System | ✗ | ✗ | ✗ | ✓ |
Your AI remembers preferences, habits, and past conversations across sessions. Tell it once, remembered forever.
AI remembers your tech stack, coding style, and project decisions. Always follows your conventions.
Bot remembers previous tickets, solutions, and customer preferences. No repeating yourself.
Add persistent memory to Claude Desktop, Cursor, or any MCP client. Zero code — just configure and go.
Install the SDK and start building with persistent memory.