Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Beyond [CLS] through Ranking by Generation (2010.03073v1)

Published 6 Oct 2020 in cs.CL and cs.IR

Abstract: Generative models for Information Retrieval, where ranking of documents is viewed as the task of generating a query from a document's LLM, were very successful in various IR tasks in the past. However, with the advent of modern deep neural networks, attention has shifted to discriminative ranking functions that model the semantic similarity of documents and queries instead. Recently, deep generative models such as GPT2 and BART have been shown to be excellent text generators, but their effectiveness as rankers have not been demonstrated yet. In this work, we revisit the generative framework for information retrieval and show that our generative approaches are as effective as state-of-the-art semantic similarity-based discriminative models for the answer selection task. Additionally, we demonstrate the effectiveness of unlikelihood losses for IR.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Cicero Nogueira dos Santos (31 papers)
  2. Xiaofei Ma (31 papers)
  3. Ramesh Nallapati (38 papers)
  4. Zhiheng Huang (33 papers)
  5. Bing Xiang (74 papers)
Citations (30)

Summary

We haven't generated a summary for this paper yet.