Papers
Topics
Authors
Recent
2000 character limit reached

SGMem: Sentence Graph Memory for Long-Term Conversational Agents (2509.21212v1)

Published 25 Sep 2025 in cs.CL and cs.IR

Abstract: Long-term conversational agents require effective memory management to handle dialogue histories that exceed the context window of LLMs. Existing methods based on fact extraction or summarization reduce redundancy but struggle to organize and retrieve relevant information across different granularities of dialogue and generated memory. We introduce SGMem (Sentence Graph Memory), which represents dialogue as sentence-level graphs within chunked units, capturing associations across turn-, round-, and session-level contexts. By combining retrieved raw dialogue with generated memory such as summaries, facts and insights, SGMem supplies LLMs with coherent and relevant context for response generation. Experiments on LongMemEval and LoCoMo show that SGMem consistently improves accuracy and outperforms strong baselines in long-term conversational question answering.

Summary

  • The paper presents SGMem, a novel framework that segments dialogue into sentences and constructs a graph for precise memory management.
  • It utilizes Sentence-BERT embeddings and a multi-hop retrieval strategy to effectively index and access dialogue segments, outperforming traditional methods on benchmarks.
  • The framework is lightweight and scalable, mitigating memory fragmentation and enabling coherent, context-aware responses in real-time multi-turn conversations.

SGMem: Sentence Graph Memory for Long-Term Conversational Agents

This paper, titled "SGMem: Sentence Graph Memory for Long-Term Conversational Agents," introduces a memory management framework specifically designed to address the challenges faced by long-term conversational agents. The proposed framework, Sentence Graph Memory (SGMem), organizes dialogue at the sentence level to effectively manage and retrieve memory across multi-turn conversations.

Introduction

Long-term conversational agents often grapple with the challenge of memory overload, as the dialogue history accumulates beyond the context window size of LLMs. Existing memory management methods focus on reducing redundancy through summarization or fact extraction but often fail to organize or retrieve information in a coherent manner across different dialogue granularities. SGMem addresses these limitations by representing dialogue as sentence-level graphs, enabling more precise retrieval of relevant memory segments.

Methodology

SGMem consists of two primary components: SGMem Construction {content} Management and SGMem Usage.

  1. SGMem Construction {content} Management: This component involves segmenting dialogue into sentences and constructing a memory graph with sentences as nodes. Sentence vectors are indexed using embeddings from models like Sentence-BERT. Edges in the graph are defined based on sentence similarity and chunk membership, allowing for efficient retrieval.
  2. SGMem Usage: In this phase, memory and sentences relevant to the input query are retrieved from the indexed graph. A multi-hop retrieval strategy refines the selection of memory chunks, integrating sentence-level context with generated memory (i.e., summaries, facts, insights). This enriches the response generation process by providing a coherent dialogue history. Figure 1

    Figure 1: Overview of the proposed Sentence Graph Memory (SGMem) framework, consisting of (a) SGMem Construction {content} Management for building sentence-level memory graphs and (b) SGMem Usage for retrieving relevant memory in long-term conversational QA.

Experimental Results

SGMem was evaluated on LongMemEval and LoCoMo datasets, specifically designed to benchmark long-term conversational agents. The SGMem variants consistently outperformed existing RAG implementations and memory management methods across diverse query types.

In particular, SGMem achieved superior accuracy in multi-session and temporally dynamic settings, demonstrating its capacity to mitigate memory fragmentation and enhance retrieval efficacy. Figure 2

Figure 2

Figure 2

Figure 2

Figure 2: Performance comparison of RAG and SGMem variants on LongMemEval and LoCoMo under Top-5 and Top-10 settings. Turn-, round-, and session-level denote raw dialogue units; M, F, I denote summary, fact, and insight, respectively.

Discussion

The SGMem framework offers several advantages over traditional methods by maintaining a detailed and fine-grained memory representation. The granularity allows for more accurate information retrieval and subsequently more accurate and contextually appropriate responses from conversational agents. One key benefit of SGMem is that it does not require complex LLM-based extractions, making it lightweight and scalable for real-time deployment.

Conclusion

SGMem represents a significant advancement in memory management for long-term conversational agents. By adopting a sentence-level graph-based approach, SGMem overcomes issues of memory fragmentation and aligns raw dialogue history with generated memory to provide coherent conversational context. The improvements observed across various query types underline the practical value of this framework in enhancing conversational agents' accuracy and coherence. Further work could involve optimizing graph construction and exploring broader memory types for more complex conversational scenarios.

In summary, SGMem introduces a practical and efficient strategy for organizing and utilizing conversational memory, with promising implications for the future design of intelligent conversational systems.

Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Youtube Logo Streamline Icon: https://streamlinehq.com