Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

MultiReQA: A Cross-Domain Evaluation for Retrieval Question Answering Models (2005.02507v1)

Published 5 May 2020 in cs.CL and cs.LG

Abstract: Retrieval question answering (ReQA) is the task of retrieving a sentence-level answer to a question from an open corpus (Ahmad et al.,2019).This paper presents MultiReQA, anew multi-domain ReQA evaluation suite com-posed of eight retrieval QA tasks drawn from publicly available QA datasets. We provide the first systematic retrieval based evaluation over these datasets using two supervised neural models, based on fine-tuning BERT andUSE-QA models respectively, as well as a surprisingly strong information retrieval baseline,BM25. Five of these tasks contain both train-ing and test data, while three contain test data only. Performance on the five tasks with train-ing data shows that while a general model covering all domains is achievable, the best performance is often obtained by training exclusively on in-domain data.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Mandy Guo (21 papers)
  2. Yinfei Yang (73 papers)
  3. Daniel Cer (28 papers)
  4. Qinlan Shen (6 papers)
  5. Noah Constant (32 papers)
Citations (43)