Memory is a core component of modern AI agents, and now it is gaining more attention as agents tackle longer tasks and more complex environments. It is responsible for many things: it helps agents store past experiences, retrieve useful information, keep track of context, and use what happened before to make better decisions later. To better understand the current landscape, weβve compiled a list of fresh memory architectures and frameworks shaping how AI agents remember, learn, and reason over time:
Agentic Memory (AgeMem)
This framework unifies short-term memory (STM) and long-term memory (LTM) inside the agent itself, so a memory management becomes part of the agentβs decision-making process. Agents identify what to store, retrieve, summarize, or discard. Plus, training with reinforcement learning improves performance and memory efficiency on long tasks. β Read more
Memex
An indexed experience memory mechanism that stores full interactions in an external memory database and keeps only compact summaries and indices in context. The agent can retrieve exact past information when needed. This improves long-horizon reasoning while keeping context small. β Read more
MemRL
Helps AI agents improve over time using episodic memory instead of retraining. The system stores past experiences and learns which strategies work best through reinforcement learning. This way, MemRL separates stable reasoning from flexible memory and lets agents adapt and get better without updating model weights. β Read more
UMA (Unified Memory Agent)
It is an RL-trained agent that actively manages its memory while answering questions. It uses a dual memory system: a compact global summary plus a structured keyβvalue Memory Bank that supports CRUD operations (create, update, delete, reorganize). It has improved long-horizon reasoning and state tracking. β Read more
Pancake
A high-performance hierarchical memory system for LLM agents that speeds up large-scale vector memory retrieval. It combines 3 techniques: 1) multi-level index caching (to exploit access patterns), 2) a hybrid graph index shared across multiple agents, and 3) coordinated GPUβCPU execution for fast updates and search. β Read more
Conditional memory
A model/agent selectively looks up stored knowledge during inference instead of activating everything. This is implemented with techniques like sparse memory tables (e.g., Engram N-gram lookup), keyβvalue memory slots, routing/gating networks that decide when to query memory, and hashed indexing for O(1) retrieval. This lets agents access specific knowledge cheaply without increasing model size or context. β Read more in our article
Multi-Agent Memory from a Computer Architecture Perspective
A short but interesting paper that envisions memory for multi-agent LLM systems as a computer architecture. It introduces ideas such as shared vs. distributed memory, a three-layer memory hierarchy (I/O, cache, memory), highlights missing protocols for cache sharing and memory access between agents, and emphasizes memory consistency as a key challenge. β Read more
