Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

In Defense of Cross-Encoders for Zero-Shot Retrieval (2212.06121v1)

Published 12 Dec 2022 in cs.IR and cs.CL

Abstract: Bi-encoders and cross-encoders are widely used in many state-of-the-art retrieval pipelines. In this work we study the generalization ability of these two types of architectures on a wide range of parameter count on both in-domain and out-of-domain scenarios. We find that the number of parameters and early query-document interactions of cross-encoders play a significant role in the generalization ability of retrieval models. Our experiments show that increasing model size results in marginal gains on in-domain test sets, but much larger gains in new domains never seen during fine-tuning. Furthermore, we show that cross-encoders largely outperform bi-encoders of similar size in several tasks. In the BEIR benchmark, our largest cross-encoder surpasses a state-of-the-art bi-encoder by more than 4 average points. Finally, we show that using bi-encoders as first-stage retrievers provides no gains in comparison to a simpler retriever such as BM25 on out-of-domain tasks. The code is available at https://github.com/guilhermemr04/scaling-zero-shot-retrieval.git

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Guilherme Rosa (3 papers)
  2. Luiz Bonifacio (9 papers)
  3. Vitor Jeronymo (11 papers)
  4. Hugo Abonizio (12 papers)
  5. Marzieh Fadaee (40 papers)
  6. Roberto Lotufo (41 papers)
  7. Rodrigo Nogueira (70 papers)
Citations (16)