Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Conformer-Kernel with Query Term Independence for Document Retrieval (2007.10434v1)

Published 20 Jul 2020 in cs.IR, cs.CL, and cs.LG

Abstract: The Transformer-Kernel (TK) model has demonstrated strong reranking performance on the TREC Deep Learning benchmark---and can be considered to be an efficient (but slightly less effective) alternative to BERT-based ranking models. In this work, we extend the TK architecture to the full retrieval setting by incorporating the query term independence assumption. Furthermore, to reduce the memory complexity of the Transformer layers with respect to the input sequence length, we propose a new Conformer layer. We show that the Conformer's GPU memory requirement scales linearly with input sequence length, making it a more viable option when ranking long documents. Finally, we demonstrate that incorporating explicit term matching signal into the model can be particularly useful in the full retrieval setting. We present preliminary results from our work in this paper.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Bhaskar Mitra (78 papers)
  2. Hamed Zamani (88 papers)
  3. Nick Craswell (51 papers)
  4. Sebastian Hofstatter (4 papers)
Citations (21)

Summary

We haven't generated a summary for this paper yet.