Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
98 tokens/sec
GPT-4o
61 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Long Time No See! Open-Domain Conversation with Long-Term Persona Memory (2203.05797v2)

Published 11 Mar 2022 in cs.CL

Abstract: Most of the open-domain dialogue models tend to perform poorly in the setting of long-term human-bot conversations. The possible reason is that they lack the capability of understanding and memorizing long-term dialogue history information. To address this issue, we present a novel task of Long-term Memory Conversation (LeMon) and then build a new dialogue dataset DuLeMon and a dialogue generation framework with Long-Term Memory (LTM) mechanism (called PLATO-LTM). This LTM mechanism enables our system to accurately extract and continuously update long-term persona memory without requiring multiple-session dialogue datasets for model training. To our knowledge, this is the first attempt to conduct real-time dynamic management of persona information of both parties, including the user and the bot. Results on DuLeMon indicate that PLATO-LTM can significantly outperform baselines in terms of long-term dialogue consistency, leading to better dialogue engagingness.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Xinchao Xu (5 papers)
  2. Zhibin Gou (15 papers)
  3. Wenquan Wu (12 papers)
  4. Zheng-Yu Niu (10 papers)
  5. Hua Wu (191 papers)
  6. Haifeng Wang (194 papers)
  7. Shihang Wang (11 papers)
Citations (95)