Papers
Topics
Authors
Recent
Detailed Answer
Quick Answer
Concise responses based on abstracts only
Detailed Answer
Well-researched responses based on abstracts and relevant paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses
Gemini 2.5 Flash
Gemini 2.5 Flash 39 tok/s
Gemini 2.5 Pro 49 tok/s Pro
GPT-5 Medium 12 tok/s Pro
GPT-5 High 18 tok/s Pro
GPT-4o 91 tok/s Pro
Kimi K2 191 tok/s Pro
GPT OSS 120B 456 tok/s Pro
Claude Sonnet 4 37 tok/s Pro
2000 character limit reached

HotpotQA: Multi-Hop QA Benchmark

Updated 10 September 2025
  • HotpotQA is a large-scale multi-hop question answering benchmark featuring 112,779 Wikipedia-based Q&A pairs with detailed, sentence-level supporting evidence.
  • It employs rigorous annotation, diverse reasoning types, and multiple data splits to challenge retrieval and inference with distractor paragraphs and full wiki settings.
  • The benchmark has driven advances in explainable QA and evidence-based systems while highlighting gaps between machine and human performance.

HotpotQA is a large-scale benchmark for multi-hop question answering, designed to systematically evaluate and advance QA systems’ ability to reason over multiple relevant paragraphs, provide explicit supporting explanations, and handle diverse natural queries. HotpotQA comprises 112,779 Wikipedia-based question–answer pairs annotated with sentence-level supporting facts, with a fine-grained taxonomy of reasoning types and rigorous evaluation protocols. Its architecture, annotation methodology, and challenge settings have shaped subsequent multi-hop QA research, serving as a reference point for explainable natural language inference and evidence-based answer modeling.

1. Dataset Organization and Splits

HotpotQA is constructed for challenging multi-hop reasoning, distinguishing itself by both scale and annotation granularity. Each example consists of a question, an answer span (directly extracted from the provided Wikipedia context), and a set of supporting sentences at the sentence level. The data is split into multiple subsets spanning increasing reasoning difficulty:

Split Name Reasoning Level # Examples Usage
train-easy single-hop 18,089 training
train-medium multi-hop 56,814 training
train-hard hard multi-hop 15,661 training
dev hard multi-hop 7,405 development
test-distractor hard multi-hop 7,405 distractor test
test-fullwiki hard multi-hop 7,405 full wiki test

Each split reflects unique retrieval and reasoning challenges: the distractor setting provides the gold paragraphs along with distractors, while the full wiki setting requires retrieval from over 5 million Wikipedia paragraphs.

2. Annotation Methodology and Feature Set

HotpotQA is defined by four key features:

  • Multi-Document Reasoning: Answering requires synthesizing evidence from at least two distinct documents.
  • Natural Diversity: Questions are unconstrained by fixed schemas or knowledge bases; they are generated by leveraging the Wikipedia hyperlink graph and entity pairs.
  • Supporting Facts: Each datum includes explicit annotations for the sentences needed to answer the question, enabling strong supervision for evidence selection and explanation.
  • Factoid Comparison Questions: The dataset introduces a unique type of comparison query (e.g., “Which city has a higher population?”), sometimes necessitating arithmetic or list reasoning beyond extraction.

This design supports fine-grained evaluation (not just answer accuracy but also evidence retrieval and explanation alignment) and enables research into the structure of reasoning chains.

3. Reasoning Taxonomy and Supervised Modeling

HotpotQA’s data supports various reasoning types:

  • Type I (Bridge Entity Reasoning): Infer an intermediate entity connecting disparate pieces of evidence before resolving the main query.
  • Type II (Intersection): Identify entities via the intersection of multiple properties.
  • Type III (Relational Reasoning via Bridge): Use relational chains to connect the answer through another entity.
  • Comparison: Directly compare two entities across a shared property.

Models are constructed with multi-task architectures: a span predictor for the answer and a binary sentence classifier for supporting facts. The joint task loss combines standard span loss with binary cross-entropy over supporting fact labels. Joint metrics are computed as products of precision and recall:

P(joint)=P(ans)×P(sup),R(joint)=R(ans)×R(sup),P^{(joint)} = P^{(ans)} \times P^{(sup)}, \qquad R^{(joint)} = R^{(ans)} \times R^{(sup)},

with joint F1 derived conventionally.

4. Benchmarking, Evaluation, and Results

HotpotQA’s challenge is quantified across distractor and full wiki test settings. Baseline performance statistics (from Table 2 of the paper):

Setting Answer EM/F1 Supporting Fact EM/F1 Joint EM/F1
Distractor 44–45 / 58–59 22 / 66 11–12 / 40–41
Full wiki 25 / 34 5 / 41 2.6 / 17.8

Providing gold supporting facts during model training improves explainability and answer F1 by ≈2 points, yet a substantial gap exists between machine and human performance. Retrieval in the full wiki setting remains significantly more difficult (due to the search space explosion and distractor prevalence).

5. Significance, Use Cases, and Extensions

HotpotQA acts as both an evaluation and development corpus for multi-hop and explainable QA. Its annotated supporting facts drive advances in reasoning supervision, step-wise pipeline architectures, and chain-of-thought modeling. The design encourages research into:

  • Sequential and decompositional reasoning pipelines (e.g., QDMR decompositions to break questions into sub-steps (Wolfson et al., 2020))
  • Evidence selection algorithms and end-to-end joint models (e.g., unified passage/sentence selection with cross-level constraints (Luo et al., 2021))
  • Hard retrieval settings requiring scalable, semantically-aware retrievers and rerankers
  • Factoid comparison reasoning and arithmetic-based QA

HotpotQA’s formulation has influenced subsequent benchmarks addressing counterfactuality, temporal reasoning, and generative, non-extractive multi-hop questions.

6. Limitations and Research Directions

HotpotQA’s reliance on Wikipedia introduces coverage constraints and the potential for data contamination, especially with models pre-trained on the same corpus. It does not natively evaluate temporal reasoning, generative answer construction, or detailed step-wise reasoning chains as in newer datasets (e.g., Counterfactual (Wu et al., 19 Feb 2024), MoreHopQA (Schnitzler et al., 19 Jun 2024), ComplexTempQA (Gruber et al., 7 Jun 2024), NovelHopQA (Gupta et al., 20 May 2025)). Performance inflation due to memorized knowledge is documented in more recent works, emphasizing the need for counterfactual and process-sensitive benchmarks.

Prospective directions for research inspired by HotpotQA include integrating sub-question decomposition, optimizing retrieval at scale, and developing architectures capable of robust cross-document explanation, especially in settings with longer contexts or under generative reasoning constraints.

7. Impact and Legacy

HotpotQA established a standard for multi-hop, explainable QA datasets, combining scalable annotation, diverse question styles, and evidence-level supervision. It catalyzed research on QA system transparency, retrieval-reasoning integration, and answer justification, and remains a central resource for benchmarking multi-hop QA, step-wise inference, and explainable neural architectures. Its limitations have helped define concrete desiderata for next-generation evaluation benchmarks targeting genuine reasoning rather than memorization, delineating the frontier in automated multi-step natural language question answering.

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to HotpotQA Benchmark.