Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Systematic Investigation of Distilling Large Language Models into Cross-Encoders for Passage Re-ranking (2405.07920v2)

Published 13 May 2024 in cs.IR

Abstract: Cross-encoders distilled from LLMs are often more effective re-rankers than cross-encoders fine-tuned on manually labeled data. However, the distilled models usually do not reach their teacher LLM's effectiveness. To investigate whether best practices for fine-tuning cross-encoders on manually labeled data (e.g., hard-negative sampling, deep sampling, and listwise loss functions) can help to improve LLM ranker distillation, we construct and release a new distillation dataset: Rank-DistiLLM. In our experiments, cross-encoders trained on Rank-DistiLLM reach the effectiveness of LLMs while being orders of magnitude more efficient. Our code and data is available at https://github.com/webis-de/msmarco-LLM-distillation.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Ferdinand Schlatt (5 papers)
  2. Maik Fröbe (20 papers)
  3. Harrisen Scells (22 papers)
  4. Shengyao Zhuang (42 papers)
  5. Bevan Koopman (37 papers)
  6. Guido Zuccon (73 papers)
  7. Benno Stein (44 papers)
  8. Martin Potthast (64 papers)
  9. Matthias Hagen (33 papers)
Citations (3)

Summary

We haven't generated a summary for this paper yet.