Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learning Diverse Document Representations with Deep Query Interactions for Dense Retrieval (2208.04232v1)

Published 8 Aug 2022 in cs.IR and cs.CL

Abstract: In this paper, we propose a new dense retrieval model which learns diverse document representations with deep query interactions. Our model encodes each document with a set of generated pseudo-queries to get query-informed, multi-view document representations. It not only enjoys high inference efficiency like the vanilla dual-encoder models, but also enables deep query-document interactions in document encoding and provides multi-faceted representations to better match different queries. Experiments on several benchmarks demonstrate the effectiveness of the proposed method, out-performing strong dual encoder baselines.The code is available at \url{https://github.com/jordane95/dual-cross-encoder

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Zehan Li (26 papers)
  2. Nan Yang (182 papers)
  3. Liang Wang (512 papers)
  4. Furu Wei (291 papers)
Citations (8)
Github Logo Streamline Icon: https://streamlinehq.com