Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

AmbigQA: Answering Ambiguous Open-domain Questions (2004.10645v2)

Published 22 Apr 2020 in cs.CL and cs.AI

Abstract: Ambiguity is inherent to open-domain question answering; especially when exploring new topics, it can be difficult to ask questions that have a single, unambiguous answer. In this paper, we introduce AmbigQA, a new open-domain question answering task which involves finding every plausible answer, and then rewriting the question for each one to resolve the ambiguity. To study this task, we construct AmbigNQ, a dataset covering 14,042 questions from NQ-open, an existing open-domain QA benchmark. We find that over half of the questions in NQ-open are ambiguous, with diverse sources of ambiguity such as event and entity references. We also present strong baseline models for AmbigQA which we show benefit from weakly supervised learning that incorporates NQ-open, strongly suggesting our new task and data will support significant future research effort. Our data and baselines are available at https://nlp.cs.washington.edu/ambigqa.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Sewon Min (45 papers)
  2. Julian Michael (28 papers)
  3. Hannaneh Hajishirzi (176 papers)
  4. Luke Zettlemoyer (225 papers)
Citations (253)

Summary

Overview of "AmbigQA: Answering Ambiguous Open-domain Questions"

The paper "AmbigQA: Answering Ambiguous Open-domain Questions" introduces an innovative task in the domain of natural language processing focused on addressing the challenges posed by ambiguity in open-domain question answering (ODQA). The authors, Min et al., propose AmbigQA, a task designed to systematically identify and resolve ambiguous questions by generating all plausible answers along with corresponding disambiguated rewrites of the original question. This contribution is underscored by the introduction of a new dataset, AmbigNQ, consisting of 14,042 questions derived from the NQ-open dataset.

Ambiguity in Open-domain Question Answering

The core problem addressed in this work is the ambiguity inherent in many natural language questions, especially in open-domain settings where questions can refer to multiple events, entities, or temporal contexts. This ambiguity often results in multiple valid answers, posing challenges for current ODQA systems that typically provide a singular response. The AmbigQA task explicitly targets this issue by requiring systems to output a set of plausible answers and reformulate the question for each answer to remove its ambiguity.

Dataset and Task Formulation

AmbigNQ, the dataset developed for this paper, is curated from the NQ-open dataset and meticulously annotated to reflect the multiplicity of potential answers for ambiguous questions. The authors demonstrate that over 50% of the questions in NQ-open exhibit some form of ambiguity, necessitating a comprehensive approach to ambiguity resolution. AmbigQA requires a two-step process: generating multiple candidate answers and producing specific question rewrites that clarify the context for each answer.

Baseline Models and Experiments

To establish a foundation for future research, the paper introduces baseline models adapted for the AmbigQA task. These models leverage weakly supervised learning techniques that incorporate the abundant data from NQ-open to enhance performance. Experimental results highlight that these baseline models effectively utilize weakly supervised approaches, showing promise in tackling the ambiguity problem specified by AmbigQA.

Implications and Future Directions

The implications of this research extend both practically and theoretically. Practically, resolving ambiguity in ODQA can lead to the development of more robust question-answering systems capable of handling the nuances of human communication more effectively. Theoretically, the work challenges the community to reevaluate traditional approaches to ODQA, encouraging the development of models that not only predict answers but also engage in deeper linguistic analysis to provide context-specific answers.

Looking forward, several avenues emerge for future exploration: the refinement of models that can seamlessly integrate answer generation and question rewriting, and the exploration of transfer learning techniques that could leverage AmbigNQ annotations to improve other NLP tasks. Furthermore, the dataset and task proposed by Min et al. are poised to drive significant research aimed at overcoming ambiguity, thus advancing the field of NLP toward achieving more nuanced understanding and generation of human language.

The authors provide access to the AmbigQA dataset and associated baselines, fostering an open research environment and inviting further advances in this burgeoning area. For researchers looking to contribute, AmbigQA offers a rich ground for experimentation and innovation in tackling ambiguity in question answering systems.