Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

ReFIT: Relevance Feedback from a Reranker during Inference (2305.11744v2)

Published 19 May 2023 in cs.IR and cs.CL

Abstract: Retrieve-and-rerank is a prevalent framework in neural information retrieval, wherein a bi-encoder network initially retrieves a pre-defined number of candidates (e.g., K=100), which are then reranked by a more powerful cross-encoder model. While the reranker often yields improved candidate scores compared to the retriever, its scope is confined to only the top K retrieved candidates. As a result, the reranker cannot improve retrieval performance in terms of Recall@K. In this work, we propose to leverage the reranker to improve recall by making it provide relevance feedback to the retriever at inference time. Specifically, given a test instance during inference, we distill the reranker's predictions for that instance into the retriever's query representation using a lightweight update mechanism. The aim of the distillation loss is to align the retriever's candidate scores more closely with those produced by the reranker. The algorithm then proceeds by executing a second retrieval step using the updated query vector. We empirically demonstrate that this method, applicable to various retrieve-and-rerank frameworks, substantially enhances retrieval recall across multiple domains, languages, and modalities.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Revanth Gangi Reddy (25 papers)
  2. Pradeep Dasigi (29 papers)
  3. Md Arafat Sultan (25 papers)
  4. Arman Cohan (121 papers)
  5. Avirup Sil (45 papers)
  6. Heng Ji (266 papers)
  7. Hannaneh Hajishirzi (176 papers)
Citations (1)
X Twitter Logo Streamline Icon: https://streamlinehq.com