Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Question Rewriting for Conversational Question Answering (2004.14652v3)

Published 30 Apr 2020 in cs.IR and cs.LG

Abstract: Conversational question answering (QA) requires the ability to correctly interpret a question in the context of previous conversation turns. We address the conversational QA task by decomposing it into question rewriting and question answering subtasks. The question rewriting (QR) subtask is specifically designed to reformulate ambiguous questions, which depend on the conversational context, into unambiguous questions that can be correctly interpreted outside of the conversational context. We introduce a conversational QA architecture that sets the new state of the art on the TREC CAsT 2019 passage retrieval dataset. Moreover, we show that the same QR model improves QA performance on the QuAC dataset with respect to answer span extraction, which is the next step in QA after passage retrieval. Our evaluation results indicate that the QR model we proposed achieves near human-level performance on both datasets and the gap in performance on the end-to-end conversational QA task is attributed mostly to the errors in QA.

An Academic Overview of "Question Rewriting for Conversational Question Answering"

In the paper "Question Rewriting for Conversational Question Answering," the authors address the challenges faced by conversational question-answering (QA) systems, focusing on maintaining context across multiple user interactions. This paper proposes a novel architecture that deconstructs the task into question rewriting (QR) and question-answering components, aiming to enhance the accuracy and effectiveness of conversational QA systems.

The central proposition of the paper is the use of a QR model to reformulate ambiguous follow-up questions into explicit questions that can be effectively interpreted by standard QA models. The authors argue that this two-step approach not only simplifies the problem by reducing it to a standard QA task but also offers additional benefits such as modularity, reusability, and the separation of error sources.

Key Contributions and Methodology

There are several key contributions in the paper:

  1. State-of-the-Art Performance: The paper introduces a QR model that achieves state-of-the-art results on the TREC Conversational Assistance Track (CAsT) dataset. The model also enhances performance on the QuAC dataset concerning answer span extraction, evidencing its utility across different types of QA tasks.
  2. Versatile QR Architecture: The QR model, based on a Transformer Decoder, utilizes the architecture of a pre-trained GPT-2 model to generate unambiguous questions from a conversational context. This approach is compared to other rewriting solutions, such as co-reference resolution and various sequence-to-sequence models, and demonstrates superior performance.
  3. Benchmark Evaluation: The authors conduct extensive evaluations using established metrics such as ROUGE for QR performance, MAP, MRR, and NDCG@3 for retrieval QA, and F1 and Exact Match (EM) for extractive QA. The QR model consistently outperforms baselines that incorporate previous conversation turns directly into the QA model without rewriting.

Analytical Results

The paper presents an insightful analysis of the error sources within the QA task, leveraging the modularity of their proposed framework. The paper highlights that a significant portion of errors arises from the QA component, even when the questions are well-formulated by humans, suggesting potential areas for further optimization in QA algorithms.

Moreover, the authors validate the value of non-conversational pre-training (such as on the MultiQA dataset), noting its potential to enhance performance significantly when coupled with fine-tuning on conversational datasets like CANARD. This underscores the adaptability and utility of the QR model across varying contexts and domains.

Implications and Future Directions

This research has important implications for the advancement of conversational AI. By decoupling the task into distinct rewriting and answering phases, the paper offers a scalable and efficient solution that can be integrated with existing QA systems seamlessly. Practically, this means that sophisticated conversational agents can be developed with increased accuracy and reduced computational overhead, given that existing QA models can be repurposed for use with minimal modifications.

The authors also provide a direction for future research, suggesting the exploration of integrated models that combine passage retrieval and answer span extraction stages. Additionally, extending QR architectures to incorporate multi-modal data or user models could enhance the interpretative capacity of conversational systems, further advancing the field of dialog systems in dynamic and context-rich environments.

In conclusion, this paper provides a substantial contribution to the field of conversational question-answering, detailing both a methodological innovation and a rigorous empirical validation of its efficacy. The approach advances the dialog between user interaction history and machine understanding, setting a clear trajectory for future developments in AI-driven conversational tools.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Svitlana Vakulenko (31 papers)
  2. Shayne Longpre (49 papers)
  3. Zhucheng Tu (11 papers)
  4. Raviteja Anantha (13 papers)
Citations (162)
Youtube Logo Streamline Icon: https://streamlinehq.com