Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

RECOMP: Improving Retrieval-Augmented LMs with Compression and Selective Augmentation (2310.04408v1)

Published 6 Oct 2023 in cs.CL
RECOMP: Improving Retrieval-Augmented LMs with Compression and Selective Augmentation

Abstract: Retrieving documents and prepending them in-context at inference time improves performance of LLM (LMs) on a wide range of tasks. However, these documents, often spanning hundreds of words, make inference substantially more expensive. We propose compressing the retrieved documents into textual summaries prior to in-context integration. This not only reduces the computational costs but also relieves the burden of LMs to identify relevant information in long retrieved documents. We present two compressors -- an extractive compressor which selects useful sentences from retrieved documents and an abstractive compressor which generates summaries by synthesizing information from multiple documents. Both compressors are trained to improve LMs' performance on end tasks when the generated summaries are prepended to the LMs' input, while keeping the summary concise.If the retrieved documents are irrelevant to the input or offer no additional information to LM, our compressor can return an empty string, implementing selective augmentation.We evaluate our approach on LLMing task and open domain question answering task. We achieve a compression rate of as low as 6% with minimal loss in performance for both tasks, significantly outperforming the off-the-shelf summarization models. We show that our compressors trained for one LM can transfer to other LMs on the LLMing task and provide summaries largely faithful to the retrieved documents.

Improving Retrieval-Augmented LLMs with Compression: A Summary

The paper "RECOMP: Improving Retrieval-Augmented LMs with Compression and Selective Augmentation" addresses a key challenge in the domain of retrieval-augmented LLMs (RALMs): the computational inefficiencies of handling long retrieved documents in the context of LLMing and open-domain question answering tasks. The authors propose a strategy to compress these retrieved documents into concise textual summaries, which can be integrated into the LLMs, thereby reducing computational burdens while maintaining or even enhancing task performance.

Compression Methodologies

The paper introduces two primary compression mechanisms:

  1. Extractive Compression: This method involves selecting the most relevant sentences from the retrieved documents, offering a sentence-level extraction that aims to be both concise and effective. The extractive compressor uses a dual encoder model optimized through contrastive learning to rank sentences that improve the target LLM’s (LM) performance.
  2. Abstractive Compression: Here, the compression is more synthesis-oriented, generating summaries that blend information from multiple documents. This method employs an encoder-decoder model, distilled from extreme-scale LLMs like GPT-3, focused on outputting summaries that not only capture relevant data but also simplify computational needs.

In both cases, the compressors perform selective augmentation, where they may return an empty string if the retrieved information is deemed irrelevant or unhelpful, effectively negating unnecessary additions to the input context.

Evaluation and Results

The approach was tested across multiple data sets, including WikiText-103 for LLMing and diverse question-answering datasets like Natural Questions (NQ), TriviaQA, and HotpotQA. Notably, the compression method achieved compression rates as low as 6% without significant performance loss. This is particularly compelling as it outperformed existing off-the-shelf summarization models in both efficiency and task accuracy.

The compressors were also tested for their ability to generalize across different languages models. Initially trained for one LM (e.g., GPT2), these compressors were able to effectively transfer to others such as GPT2-XL and GPT-J, demonstrating the robustness of the textual summarization approach.

Implications and Future Directions

The findings in this paper underline the potential of combining retrieval-augmented methods with informed compression strategies to enhance the application of LMs in real-world scenarios. By reducing the amount of information LLMs need to process, these methods can lead to significant computational savings and expand the usability of LMs in environments with limited processing capabilities.

Future research could further refine these compressors, particularly focusing on achieving adaptive augmentation, where the number of prepended sentences is dynamically optimized on a per-task basis. Moreover, improving the faithfulness and comprehensiveness of the abstractive summaries remains an open challenge, especially for tasks requiring complex multi-hop reasoning like HotpotQA.

This research proposes a promising step in optimizing RALMs, with potential applications in more efficient document processing, scalable deployment of LMs in constrained environments, and advancing the state-of-the-art in AI-driven language processing tasks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Fangyuan Xu (10 papers)
  2. Weijia Shi (55 papers)
  3. Eunsol Choi (76 papers)
Citations (103)
Youtube Logo Streamline Icon: https://streamlinehq.com