Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Unraveling and Mitigating Retriever Inconsistencies in Retrieval-Augmented Large Language Models (2405.20680v4)

Published 31 May 2024 in cs.AI and cs.CL

Abstract: Although Retrieval-Augmented LLMs (RALMs) demonstrate their superiority in terms of factuality, they do not consistently outperform the original retrieval-free LLMs (LMs). Our experiments reveal that this example-level performance inconsistency exists not only between retrieval-augmented and retrieval-free LM but also among different retrievers. To understand this phenomenon, we investigate the degeneration behavior of RALMs and theoretically decompose it into four categories. Further analysis based on our decomposition reveals that the innate difference in knowledge sources and the unpredictable degeneration of the reader model contribute most to the inconsistency. Drawing from our analysis, we introduce Ensemble of Retrievers (EoR), a trainable framework that can adaptively retrieve from different knowledge sources and effectively decrease unpredictable reader errors. Our experiments on Open Domain Question Answering show that EoR substantially improves performance over the RALM with a single retriever by considerably reducing inconsistent behaviors.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Mingda Li (95 papers)
  2. Xinyu Li (136 papers)
  3. Yifan Chen (164 papers)
  4. Wenfeng Xuan (1 paper)
  5. Weinan Zhang (322 papers)
Citations (1)