Memory is a core component of modern AI agents, and now it is gaining more attention as agents tackle longer tasks and more complex environments. It is responsible for many things: it helps agents store past experiences, retrieve useful information, keep track of context, and use what happened before to make better decisions later. To better understand the current landscape, we’ve compiled a list of fresh memory architectures and frameworks shaping how AI agents remember, learn, and reason over time:

  1. Agentic Memory (AgeMem)

    This framework unifies short-term memory (STM) and long-term memory (LTM) inside the agent itself, so a memory management becomes part of the agent’s decision-making process. Agents identify what to store, retrieve, summarize, or discard. Plus, training with reinforcement learning improves performance and memory efficiency on long tasks. β†’ Read more

  2. Memex

    An indexed experience memory mechanism that stores full interactions in an external memory database and keeps only compact summaries and indices in context. The agent can retrieve exact past information when needed. This improves long-horizon reasoning while keeping context small. β†’ Read more

  3. MemRL

    Helps AI agents improve over time using episodic memory instead of retraining. The system stores past experiences and learns which strategies work best through reinforcement learning. This way, MemRL separates stable reasoning from flexible memory and lets agents adapt and get better without updating model weights. β†’ Read more

  4. UMA (Unified Memory Agent)

    It is an RL-trained agent that actively manages its memory while answering questions. It uses a dual memory system: a compact global summary plus a structured key–value Memory Bank that supports CRUD operations (create, update, delete, reorganize). It has improved long-horizon reasoning and state tracking. β†’ Read more

  5. Pancake

    A high-performance hierarchical memory system for LLM agents that speeds up large-scale vector memory retrieval. It combines 3 techniques: 1) multi-level index caching (to exploit access patterns), 2) a hybrid graph index shared across multiple agents, and 3) coordinated GPU–CPU execution for fast updates and search. β†’ Read more

  6. Conditional memory

    A model/agent selectively looks up stored knowledge during inference instead of activating everything. This is implemented with techniques like sparse memory tables (e.g., Engram N-gram lookup), key–value memory slots, routing/gating networks that decide when to query memory, and hashed indexing for O(1) retrieval. This lets agents access specific knowledge cheaply without increasing model size or context. β†’ Read more in our article

  7. Multi-Agent Memory from a Computer Architecture Perspective

    A short but interesting paper that envisions memory for multi-agent LLM systems as a computer architecture. It introduces ideas such as shared vs. distributed memory, a three-layer memory hierarchy (I/O, cache, memory), highlights missing protocols for cache sharing and memory access between agents, and emphasizes memory consistency as a key challenge. β†’ Read more

Also, subscribe to our X, Threads and BlueSky

to get unique content on every social media

Reply

Avatar

or to participate

Keep Reading