Papers
Topics
Authors
Recent
2000 character limit reached

MemoryAgentBench: LLM Memory Benchmark

Updated 22 December 2025
  • MemoryAgentBench is a unified benchmark suite designed to assess LLM agents' memory through accurate retrieval, test-time learning, long-range understanding, and conflict resolution.
  • It employs incremental chunking and multi-competency sampling to simulate realistic multi-turn and multi-session user-agent interactions using standardized protocols.
  • Evaluations reveal trade-offs among memory architectures, highlighting challenges in conflict resolution and the need for hybrid approaches to improve efficiency and scalability.

MemoryAgentBench is a unified benchmark suite for evaluating the long-term, interactive, and adaptive memory capabilities of LLM agents. MemoryAgentBench specifically targets the competencies of information retention, update, retrieval, and conflict resolution, as encountered in realistic multi-turn or multi-session user-agent interactions. The benchmark addresses the insufficiencies of prior approaches, which primarily focus on static reading comprehension or isolated document-based QA tasks, by introducing datasets, metrics, and protocols that operationalize interactive memory management and agent learning from ongoing information streams (Hu et al., 7 Jul 2025).

1. Motivation and Scope

MemoryAgentBench was developed in response to major gaps in existing memory benchmarks for LLM agents. Conventional testbeds focus primarily on either reasoning, tool orchestration, or “needle-in-a-haystack” retrieval within static, ultra-long contexts. These are insufficient for evaluating memory agents, which must incrementally accumulate, retrieve, and update knowledge across extended dialogues and evolving contexts.

Key limitations of previous approaches include:

  • Emphasis on monolithic documents rather than incremental, multi-turn input streams.
  • Evaluation limited to single competencies (e.g., retrieval) rather than the full spectrum of memory functions.
  • Lack of systematic testing for critical abilities such as test-time learning (incorporation of user-taught rules) and conflict resolution (handling of contradictory updates).

MemoryAgentBench situates memory as a multi-dimensional faculty, essential for the long-term viability of LLM-based agents deployed in dynamic environments (Hu et al., 7 Jul 2025).

2. Memory Competencies and Task Taxonomy

MemoryAgentBench operationalizes memory quality along four axes, each corresponding to core agentic requirements:

  1. Accurate Retrieval (AR): The ability to locate and extract precise information from arbitrarily long, sequential interaction histories—akin to multi-hop “needle-in-a-haystack” retrieval.
  2. Test-Time Learning (TTL): Ingesting and reliably applying new classification labels, rules, or procedures provided by the user during dialogue, without explicit parameter updates.
  3. Long-Range Understanding (LRU): Forming coherent, abstract summaries or global views over extended narratives or multi-domain dialogues (e.g., summarizing a full novel-length conversation).
  4. Conflict Resolution (CR): Detecting and overwriting outdated facts when new, potentially contradictory information is introduced, ensuring subsequent queries only reflect the newest valid data.

Each competency is mapped to carefully constructed datasets and evaluation metrics:

Competency Dataset Examples Evaluation Metric(s)
Accurate Retrieval RULER-QA, NIAH-MQ, ∞Bench-QA Substring Exact Match, Recall
Test-Time Learning BANKING-77, CLINC150, Redial Classification acc., Recall@5
Long-Range Understanding ∞Bench-Sum GPT-4–scored F1 (summary)
Conflict Resolution FactConsolidation-SH/MH Latest-fact Exact Match

This taxonomy ensures comprehensive coverage, distinguishing MemoryAgentBench from narrowly-scoped, retrieval-only frameworks (Hu et al., 7 Jul 2025).

3. Dataset Construction and Protocol

MemoryAgentBench assembles and reformulates both synthetic and naturalistic datasets to systematically probe each competency. Key dataset construction strategies include:

  • Incremental Chunking: Documents and interaction histories are split into sequential 512- or 4,096-token chunks, simulating agents that process and memorize inputs over time.
  • Multi-Competency Sampling: For AR, existing “needle-in-a-haystack” and reading comprehension tasks are restructured to require chunk-by-chunk memory accumulation and delayed retrieval. TTL is emulated via concatenation of large numbers of labeled examples and agent classification of novel items after reading the entire set. LRU is instantiated through summarization queries on 172 K+ token extended narratives. CR benchmarks leverage concatenated edit pairs and require the agent to answer based only on the latest fact in the history.
  • Uniform Task Templates: Construction prompts are standardized, with separate memory-construction and query phases. Ablations alter chunk sizes, retriever TopK, and session lengths to probe scaling and edge cases.

The protocol requires agents to build memory from input streams before answering held-out questions, emulating realistic agent deployment (Hu et al., 7 Jul 2025).

4. Agent Architectures and Implementation

The benchmark assesses a diversity of memory handling paradigms, all typically configured with 128 K–200 K context windows when models support such length:

  • Long-Context Agents: Pure FIFO buffers retaining up to 128 K/200 K tokens, with oldest content evicted as new tokens arrive. Instantiated with models such as GPT-4o, Gemini-2.0-Flash, and Claude-3.7-Sonnet.
  • Retrieval-Augmented Generation (RAG) Agents:
    • BM25-based sparse retrievers (keyword match).
    • Embedding-based dense retrievers (e.g., NV-Embed-v2) operating over token chunks.
    • Structure-augmented RAG (RAPTOR, GraphRAG), supporting hierarchical or graph-based retrieval.
  • Agentic Memory Agents: Methods that incorporate external memory modules, chain-of-thought stores (MemGPT), or iterative retrieve→reason loops (Self-RAG).
  • Experimental Details: All agents process the incremental input stream before receiving the test questions. Ablations explore the effect of retriever TopK and chunk granularity on both performance and computational efficiency.

MemoryAgentBench allows for paradigm-agnostic comparison of both naïve and advanced memory architectures in terms of all four memory competencies (Hu et al., 7 Jul 2025).

5. Metrics and Results

Task-specific and aggregate metrics are reported:

  • Substring Exact Match (SubEM): Measures whether agent answers are exact substrings of the gold reference.
  • Recall (multi-value): Counts the fraction of ground-truth answer items successfully retrieved.
  • Classification Accuracy and Recall@5: Standard for TTL and recommendations.
  • GPT-4–Judged F1: Used for summarization outputs in LRU.
  • Conflict Resolution Accuracy (CR-Acc): Ensures only the latest, valid fact is considered in answers when contradictions exist.

Selected findings (see Table 2 in (Hu et al., 7 Jul 2025)):

Agent Class AR (%) TTL (%) LRU (%) CR-SH (%) / CR-MH (%)
Long-Context 43.5 82.0 28.9 45.0 / 5.0
Simple RAG (BM25) 61.0 75.4 20.9 56.0 / 3.0
Embedding RAG 65.0 69.4 20.7 55.0 / 6.0
Structure RAG 57.2 61.4 14.6 54.0 / 5.0
Agentic (MemGPT) 30.6 67.6 2.5 28.0 / 3.0
  • RAG methods (especially dense retrieval) outperform long-context agents on retrieval, but struggle with global summarization and in-situ test-time learning.
  • Long-context agents dominate TTL and LRU but fail as context size exceeds their internal window.
  • All paradigms exhibit dramatic failures on multi-hop conflict resolution: best accuracy remains at or below 6% for CR-MH, even in the most advanced models.
  • Latency reflects the trade-off between retrieval system complexity and in-context memory use: RAG generally incurs a per-query latency of 1–2 s (plus memory construction), long-context agents require ∼5 s per query (Hu et al., 7 Jul 2025).

6. Insights, Limitations, and Future Research

Empirical evidence from MemoryAgentBench reveals:

  • No single-memory paradigm is sufficient across all competencies. RAG offers precise retrieval; long-context windows are optimal for sequential learning and summarization; agentic/iterative paradigms remain immature.
  • Conflict resolution is a key bottleneck: Even with advanced indexing and agentic loops, handling contradicting updates and correctly overwriting outdated facts remains unsolved.
  • Scalability and efficiency: Pure in-context methods are constrained by context window size; RAG’s effectiveness depends on retriever granularity and chunking, with nontrivial latency and resource costs.
  • Towards hybrid architectures: Integrating buffer, dense-retrieval, and structured (e.g., graph or temporal) memory may offer better global-local trade-offs.
  • Real-world realism: Most benchmarks, including MemoryAgentBench, are derived from synthetic or reformulated academic datasets. Incorporating real multi-user deployments and conversational noise remains an open direction (Hu et al., 7 Jul 2025).

MemoryAgentBench establishes the foundation for systematic, multi-dimensional assessment of memory agents and provides standardized evaluation to drive further research in robust, interactive long-term memory for LLM-based agents.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Whiteboard

Topic to Video (Beta)

Follow Topic

Get notified by email when new papers are published related to MemoryAgentBench.