Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

QRMeM: Unleash the Length Limitation through Question then Reflection Memory Mechanism (2406.13167v2)

Published 19 Jun 2024 in cs.CL

Abstract: While LLMs have made notable advancements in natural language processing, they continue to struggle with processing extensive text. Memory mechanism offers a flexible solution for managing long contexts, utilizing techniques such as compression, summarization, and structuring to facilitate nuanced and efficient handling of large volumes of text. However, existing techniques face challenges with static knowledge integration, leading to insufficient adaptation to task-specific needs and missing multi-segmentation relationships, which hinders the dynamic reorganization and logical combination of relevant segments during the response process. To address these issues, we introduce a novel strategy, Question then Reflection Memory Mechanism (QRMeM), incorporating a dual-structured memory pool. This pool synergizes static textual content with structured graph guidance, fostering a reflective trial-and-error approach for navigating and identifying relevant segments. Our evaluation across multiple-choice questions (MCQ) and multi-document question answering (Multi-doc QA) benchmarks showcases QRMeM enhanced performance compared to existing approaches.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Bo Wang (823 papers)
  2. Heyan Huang (107 papers)
  3. Yixin Cao (138 papers)
  4. Jiahao Ying (14 papers)
  5. Wei Tang (135 papers)
  6. Chong Feng (11 papers)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets