Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
98 tokens/sec
GPT-4o
61 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

MemoChat: Tuning LLMs to Use Memos for Consistent Long-Range Open-Domain Conversation (2308.08239v2)

Published 16 Aug 2023 in cs.CL

Abstract: We propose MemoChat, a pipeline for refining instructions that enables LLMs to effectively employ self-composed memos for maintaining consistent long-range open-domain conversations. We demonstrate a long-range open-domain conversation through iterative "memorization-retrieval-response" cycles. This requires us to carefully design tailored tuning instructions for each distinct stage. The instructions are reconstructed from a collection of public datasets to teach the LLMs to memorize and retrieve past dialogues with structured memos, leading to enhanced consistency when participating in future conversations. We invite experts to manually annotate a test set designed to evaluate the consistency of long-range conversations questions. Experiments on three testing scenarios involving both open-source and API-accessible chatbots at scale verify the efficacy of MemoChat, which outperforms strong baselines. Our codes, data and models are available here: https://github.com/LuJunru/MemoChat.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Junru Lu (15 papers)
  2. Siyu An (9 papers)
  3. Mingbao Lin (78 papers)
  4. Gabriele Pergola (26 papers)
  5. Yulan He (113 papers)
  6. Di Yin (26 papers)
  7. Xing Sun (93 papers)
  8. Yunsheng Wu (25 papers)
Citations (22)