Agent MemoryMemory ArchitectureCognitive ArchitectureSeries

From Human Memory to Machine Memory: A Field Guide to AI Memory Architecture

Ryan Musser
Ryan Musser
Founder

Why a series about memory

For the last two years, almost every interesting paper about AI agents has had the word "memory" in its title. Vector databases. Episodic stores. Sleep-time agents. Reflection. Reconsolidation. The vocabulary keeps growing, and most of it is borrowed straight from cognitive psychology and neuroscience textbooks. That is not an accident. The brain has been working on the same problem (how to store, retrieve, and update useful information across time) for hundreds of millions of years. AI engineers are starting to notice that the answers are sitting in plain sight.

This series is a guided tour through the parallels. It maps 18 components of human memory to their closest analogs in modern AI systems and tells you, plainly, which ones are well-solved, which are still active research, and which are basically not built yet.

The four academic frameworks worth knowing

Four papers shaped the way people now talk about agent memory. You will see references to all of them throughout the series:

  • CoALA (Sumers et al., 2023) borrowed Tulving's classification (working, episodic, semantic, procedural) and used it to lay out a clean taxonomy for language agents.
  • Memory in the Age of AI Agents (Hu et al., 2025) extended that taxonomy with a Forms / Functions / Dynamics axis, arguing that the older "short-term vs long-term" split was no longer enough.
  • From Human Memory to AI Memory (Wu et al., 2025) introduced a 3D-8Q taxonomy mapping memory across object, form, and time.
  • Memp: Exploring Agent Procedural Memory (Fang et al., 2025) zoomed in on procedural memory, the least-built layer in agents today, and proposed a Build / Retrieve / Update lifecycle.

A fifth paper, HippoRAG (Gutierrez et al., NeurIPS 2024), deserves a special mention because it is the most biologically explicit system in production: it directly implements the hippocampal indexing theory using a knowledge graph as the index and an LLM as the parametric cortex. It outperforms standard RAG by up to 20% on multi-hop QA while being 10-30x cheaper.

The headline finding

Across all 18 subsystems, one pattern keeps repeating: retrieval and indexing are well-solved. Consolidation, forgetting, and metacognition are not. Modern HNSW indexes, transformer attention, and hybrid search give us mature analogs for hippocampal pattern completion and spreading activation. But the dynamic, adaptive parts of memory (sleep-dependent consolidation, emotional salience weighting, self-monitoring) are barely sketched. The gap is in knowing what to remember, when to forget, and how much to trust what you recall.

How the series is organized

Each post is a stand-alone read of about 900 words, structured the same way: the biology, the technology, then the gap. You can read them in order or jump straight to the component that is bothering you in production right now.

Short-term systems

How information enters the brain and what survives the first second:

Long-term stores

The three classic boxes from Tulving's classification, each with its own analog in production:

The memory lifecycle

What happens to a memory between the moment it is formed and the moment it is recalled (or lost):

The smart layers

Less obvious mechanisms that decide what gets remembered, how memories link to each other, and how confident the system is in its own recall:

Special-purpose systems

Two pieces that do not fit cleanly into the lifecycle but matter a lot for real agents:

Who this is for

If you are building an agent and you keep reaching for the same vector database trick to solve every memory problem, this series is for you. The brain figured out a long time ago that one mechanism cannot do all of memory's jobs. Sensory buffers, working memory, episodic stores, and procedural skills are different things for good reasons. Once you can see those reasons, the design choices in your own system get a lot clearer.

Start anywhere. The order is suggested, not required. And if you read one and want to argue, that is the point.

From Human Memory to Machine Memory: A Field Guide to AI Memory Architecture | TypeGraph