There's a failure mode in multi-agent systems that doesn't show up in benchmarks: procedural amnesia.
My self-improving agents could remember facts. They could remember context. But two weeks after successfully creating a Google Doc, writing a Day One journal entry, or making a phone call — they'd look at me blankly and say they didn't know how.
They knew what they'd done. They had no persistent memory of how.
The fix wasn't more context. It wasn't a bigger model. It was embarrassingly simple: a shared Markdown file called PLAYBOOKS.md.
The rule is one line: every agent, every time they successfully complete a non-trivial task, documents the exact steps before the session ends. Not a summary. The actual commands, the tools, the gotchas, the sequence. A recipe.
Before starting any task that sounds familiar, they read it first.
The result: agents that actually get smarter over time — not because the model improved, but because the system does. Institutional knowledge that survives session resets, model swaps, and context compaction.
This is also why fewer agents with better memory outperform a swarm of forgetful ones. Depth beats breadth when each agent retains what it learns.
The real insight: most "AI capability" problems are actually memory architecture problems. The model knows how to do the thing. The system just never wrote it down. If you're building agents that run autonomously, strategic memory architecture isn't optional — it's the foundation everything else depends on.