Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Fine-Grained Relevance Annotations for Multi-Task Document Ranking and Question Answering (2008.05363v1)

Published 12 Aug 2020 in cs.IR and cs.CL

Abstract: There are many existing retrieval and question answering datasets. However, most of them either focus on ranked list evaluation or single-candidate question answering. This divide makes it challenging to properly evaluate approaches concerned with ranking documents and providing snippets or answers for a given query. In this work, we present FiRA: a novel dataset of Fine-Grained Relevance Annotations. We extend the ranked retrieval annotations of the Deep Learning track of TREC 2019 with passage and word level graded relevance annotations for all relevant documents. We use our newly created data to study the distribution of relevance in long documents, as well as the attention of annotators to specific positions of the text. As an example, we evaluate the recently introduced TKL document ranking model. We find that although TKL exhibits state-of-the-art retrieval results for long documents, it misses many relevant passages.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Sebastian Hofstätter (31 papers)
  2. Markus Zlabinger (7 papers)
  3. Mete Sertkan (10 papers)
  4. Michael Schröder (8 papers)
  5. Allan Hanbury (45 papers)
Citations (11)

Summary

We haven't generated a summary for this paper yet.