Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

ERNIE-Search: Bridging Cross-Encoder with Dual-Encoder via Self On-the-fly Distillation for Dense Passage Retrieval (2205.09153v1)

Published 18 May 2022 in cs.CL

Abstract: Neural retrievers based on pre-trained LLMs (PLMs), such as dual-encoders, have achieved promising performance on the task of open-domain question answering (QA). Their effectiveness can further reach new state-of-the-arts by incorporating cross-architecture knowledge distillation. However, most of the existing studies just directly apply conventional distillation methods. They fail to consider the particular situation where the teacher and student have different structures. In this paper, we propose a novel distillation method that significantly advances cross-architecture distillation for dual-encoders. Our method 1) introduces a self on-the-fly distillation method that can effectively distill late interaction (i.e., ColBERT) to vanilla dual-encoder, and 2) incorporates a cascade distillation process to further improve the performance with a cross-encoder teacher. Extensive experiments are conducted to validate that our proposed solution outperforms strong baselines and establish a new state-of-the-art on open-domain QA benchmarks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (11)
  1. Yuxiang Lu (26 papers)
  2. Yiding Liu (30 papers)
  3. Jiaxiang Liu (39 papers)
  4. Yunsheng Shi (5 papers)
  5. Zhengjie Huang (25 papers)
  6. Shikun Feng Yu Sun (1 paper)
  7. Hao Tian (146 papers)
  8. Hua Wu (191 papers)
  9. Shuaiqiang Wang (68 papers)
  10. Dawei Yin (165 papers)
  11. Haifeng Wang (194 papers)
Citations (55)

Summary

We haven't generated a summary for this paper yet.