Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
51 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Enhancing Retrieval-Augmented Large Language Models with Iterative Retrieval-Generation Synergy (2305.15294v2)

Published 24 May 2023 in cs.CL
Enhancing Retrieval-Augmented Large Language Models with Iterative Retrieval-Generation Synergy

Abstract: LLMs are powerful text processors and reasoners, but are still subject to limitations including outdated knowledge and hallucinations, which necessitates connecting them to the world. Retrieval-augmented LLMs have raised extensive attention for grounding model generation on external knowledge. However, retrievers struggle to capture relevance, especially for queries with complex information needs. Recent work has proposed to improve relevance modeling by having LLMs actively involved in retrieval, i.e., to improve retrieval with generation. In this paper, we show that strong performance can be achieved by a method we call Iter-RetGen, which synergizes retrieval and generation in an iterative manner. A model output shows what might be needed to finish a task, and thus provides an informative context for retrieving more relevant knowledge which in turn helps generate a better output in the next iteration. Compared with recent work which interleaves retrieval with generation when producing an output, Iter-RetGen processes all retrieved knowledge as a whole and largely preserves the flexibility in generation without structural constraints. We evaluate Iter-RetGen on multi-hop question answering, fact verification, and commonsense reasoning, and show that it can flexibly leverage parametric knowledge and non-parametric knowledge, and is superior to or competitive with state-of-the-art retrieval-augmented baselines while causing fewer overheads of retrieval and generation. We can further improve performance via generation-augmented retrieval adaptation.

Introduction

A new method, termed TER-RETGEN (Iterative Retrieval-Generation Synergy), has been introduced, offering a significant advance in retrieval-augmented LLMs. TER-RETGEN operates under the premise that a model's initial response can inform the retrieval of relevant information for the next iteration, thus enhancing the quality of the generated content. This iterative process allows for a dynamic interplay between retrieval and generation, potentially resolving the challenges faced by LLMs in terms of staying current and avoiding hallucinations, particularly in complex query scenarios.

Retrieval-Augmented Generation

Traditional retrieval-augmented LLMs typically rely on a single retrieval step, falling short in situations that demand nuanced and elaborate information synthesis. TER-RETGEN distinguishes itself by handling all retrieved knowledge collectively without imposing structural constraints within the generation process. This facilitates a more fluid integration of parametric knowledge (learned from training data) and non-parametric knowledge (external information), even as the complexity of information needs escalates.

Experimental Validation

TER-RETGEN has been benchmarked against several retrieval-augmented models across tasks including multi-hop question answering, fact verification, and commonsense reasoning. The results exhibited notable accuracy improvements, demonstrating TER-RETGEN's adaptability in leveraging both parametric and non-parametric knowledge sources. Furthermore, TER-RETGEN achieves these results with minimal overhead, refining the interaction between retrieval and generation components without necessitating complex processing pipelines.

Further Insights and Adaptations

One of the remarkable aspects of TER-RETGEN is its ability to use generative outputs from previous iterations to guide retrieval adaptation. By distilling knowledge from model generations into the dense retriever, TER-RETGEN not only enhances performance but also reduces the iterative load. Specifics illustrated in the paper show that integrating generation into retrieval adaptation brings considerable improvements while trimming down both the retrieval and iteration overheads, thereby offering a more streamlined approach.

In summary, TER-RETGEN emerges as a strong contender for handling complex informational needs in retrieval-augmented settings, balancing robustness with simplicity of implementation.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Zhihong Shao (20 papers)
  2. Yeyun Gong (78 papers)
  3. Yelong Shen (83 papers)
  4. Minlie Huang (225 papers)
  5. Nan Duan (172 papers)
  6. Weizhu Chen (128 papers)
Citations (145)