Open Source · Serverless · AWS-native

The memory infrastructure for AI agents

One API for four memory types. Working, semantic, episodic, and procedural — all serverless, all AWS-native. Give your agents persistent memory in minutes, not months.

View on GitHub
DynamoDBpgvectorS3Bedrock Titan
Integrations

Built for your stack

LangGraph

Drop-in CheckpointSaver

LangChain

Memory retriever + tool

CrewAI

Shared agent memory

AutoGen

Persistent state store

OpenAI Agents SDK

Function tool integration

The problem

AI agents are stateless by default

To give them memory, developers stitch together four databases, four clients, four billing accounts — with no unified query layer.

Four databases to stitch together

Redis for state. Pinecone for vectors. Postgres for structured data. S3 for logs. Four billing accounts, four clients, zero unified memory layer.

No unified query layer

State lives here. Vectors live there. Events are somewhere else. You write glue code instead of writing agents.

Agents forget everything

Without persistence, every invocation starts from scratch. Context, preferences, and history vanish between calls.

The solution

One API. Four memory types.

AWS-native serverless infrastructure. No LLM required for CRUD operations.

<10ms

Working Memory

Key-value state in DynamoDB. Sub-10ms reads with optimistic locking and configurable TTL. Ideal for agent step state and task context.

1024d

Semantic Memory

Natural-language text stored as 1024-dimensional vectors in Aurora pgvector. Auto-embedded via Bedrock Titan. Duplicates are merged, not re-inserted.

Hot + Cold

Episodic Memory

Append-only time-series event log. Hot data in DynamoDB, automatically tiered to S3. Full session replay and time-range queries.

v0.2

Procedural Memory

Tool definitions, prompt templates, schemas, and rules stored in Postgres. Version-controlled and queryable by name. Schema is live; SDK methods ship in v0.2.

Quickstart

Your first agent memory in 15 lines

Install pip install mnemora and you're ready.

quickstart.py
Python
1from mnemora import MnemoraSync
2 
3with MnemoraSync(api_key="mnm_...") as client:
4 # Store working-memory state
5 client.store_state("agent-1", {"task": "summarize Q4", "step": 1})
6 
7 # Semantic memory — auto-embedded server-side
8 client.store_memory("agent-1", "User prefers bullet points over prose.")
9 
10 # Vector search across all stored memories
11 results = client.search_memory("user formatting preferences", agent_id="agent-1")
12 for r in results:
13 print(r.content, r.similarity_score)
14 
15 # Log an episode to the time-series history
16 client.store_episode(agent_id="agent-1", session_id="sess-001",
17 type="action", content={"tool": "summarize", "input": "Q4 report"})
Comparison

How Mnemora compares

Concrete data. No hype.

FeatureMnemora
Mem0ZepLetta
Memory types4 (state, semantic, episodic, procedural)1 (semantic only)2 (semantic + temporal)2 (core + archival)
Vector searchpgvector 1024dExternal DBBuilt-inBuilt-in
LLM required for CRUDEvery opEvery op
Serverless
Self-hostablePartial
Multi-tenant
LangGraph checkpoints
State latency<10ms~500ms<200ms~1s

Data based on public documentation as of 2025. Subject to change.

Why Mnemora

Designed different by design

No LLM in Your CRUD Path

Mem0 and Letta call an LLM for every memory operation, adding latency and token cost. Mnemora does direct database CRUD. State reads are sub-10ms. No LLM overhead, no rate limits, no surprise bills from embedding calls you didn't ask for.

Serverless-first

Every component scales to zero when idle. DynamoDB on-demand, Aurora Serverless v2, Lambda, S3. You pay per request.

Multi-tenant by Default

Each API key scoped to a tenant. Data is never mixed at the database layer. Built for SaaS.

LangGraph Native

Drop-in MnemoraCheckpointSaver with optimistic locking. Zero config.

FAQ

Frequently asked questions

Blog

From the team

All posts
Guide8 min read

Understanding Agent Memory: A Developer's Guide

A deep dive into the four types of memory AI agents need and why each matters for building production-grade systems.

Coming soon
Tutorial12 min read

Building Persistent AI Agents with LangGraph and Mnemora

Step-by-step tutorial: how to use MnemoraCheckpointSaver to give your LangGraph agents durable, cross-session memory.

Coming soon
Architecture6 min read

Why Serverless Memory Matters for AI at Scale

The economics of serverless infrastructure for AI workloads — and why stateless-by-default is the wrong architecture choice.

Coming soon
Pricing

Simple, transparent pricing

Start free. Scale as you grow. No surprises.

Free

For exploration and side projects

Free
  • 500 API calls / day
  • 1 agent
  • 50 MB storage
  • 5K vectors
  • Community support
Get started

Starter

For early-stage products

$29/month
  • 5,000 API calls / day
  • 10 agents
  • 500 MB storage
  • 50K vectors
  • Email support
Subscribe
Most popular

Pro

For production applications

$49/month
  • 25,000 API calls / day
  • 50 agents
  • 5 GB storage
  • 250K vectors
  • Priority support
Subscribe

Scale

For high-volume teams

$99/month
  • 50,000 API calls / day
  • Unlimited agents
  • 10 GB storage
  • 500K vectors
  • Dedicated support
Subscribe

Enterprise

Unlimited everything, custom SLA, SSO/SAML, VPC peering, and dedicated infrastructure for organizations with custom needs.

Contact us

All plans include TLS encryption, AWS-native infrastructure, all four memory types, and the full Python SDK. No credit card required for Free.

Ready to give your agents memory?

Start in under 5 minutes. No infrastructure to configure. No servers to manage. Just memory that works.