Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Retrospective Reader for Machine Reading Comprehension (2001.09694v4)

Published 27 Jan 2020 in cs.CL, cs.AI, and cs.IR

Abstract: Machine reading comprehension (MRC) is an AI challenge that requires machine to determine the correct answers to questions based on a given passage. MRC systems must not only answer question when necessary but also distinguish when no answer is available according to the given passage and then tactfully abstain from answering. When unanswerable questions are involved in the MRC task, an essential verification module called verifier is especially required in addition to the encoder, though the latest practice on MRC modeling still most benefits from adopting well pre-trained LLMs as the encoder block by only focusing on the "reading". This paper devotes itself to exploring better verifier design for the MRC task with unanswerable questions. Inspired by how humans solve reading comprehension questions, we proposed a retrospective reader (Retro-Reader) that integrates two stages of reading and verification strategies: 1) sketchy reading that briefly investigates the overall interactions of passage and question, and yield an initial judgment; 2) intensive reading that verifies the answer and gives the final prediction. The proposed reader is evaluated on two benchmark MRC challenge datasets SQuAD2.0 and NewsQA, achieving new state-of-the-art results. Significance tests show that our model is significantly better than the strong ELECTRA and ALBERT baselines. A series of analysis is also conducted to interpret the effectiveness of the proposed reader.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Zhuosheng Zhang (125 papers)
  2. Junjie Yang (74 papers)
  3. Hai Zhao (227 papers)
Citations (223)

Summary

Retrospective Reader for Enhanced Machine Reading Comprehension

The paper "Retrospective Reader for Machine Reading Comprehension" by Zhang, Yang, and Zhao presents an innovative approach to addressing the challenges inherent in machine reading comprehension (MRC), particularly in scenarios where questions posed may be unanswerable. This is a non-trivial task within artificial intelligence that requires systems to not only derive correct answers from given passages but also effectively identify when no answer is appropriate. The authors propose a novel framework, termed the Retro-Reader, that is specifically designed to improve performance in these MRC tasks.

Theoretical and Practical Advancements

The Retro-Reader introduces a dual-stage reading and verification strategy inspired by human cognitive processes during reading comprehension. This involves:

  1. Sketchy Reading: An initial pass that broadly assesses the interplay between the question and passage, forming a preliminary judgment on the answerability of the question.
  2. Intensive Reading: A more detailed examination that verifies the initial judgment and refines or confirms the final prediction.

This framework addresses two primary aspects of the MRC task with unanswerable questions: accurately providing answers when possible, and effectively identifying when a question cannot be answered given the context. The inclusion of a dedicated verifier module, split into external and internal components, allows for a more nuanced assessment of answerability and adds depth to the Retro-Reader's capabilities beyond existing MRC models that rely heavily on pre-trained LLMs (PrLMs).

Empirical Evaluation and Results

The proposed Retro-Reader was empirically tested on two key benchmark datasets: SQuAD2.0 and NewsQA. These datasets are well-known in the community for presenting both answerable and unanswerable questions, thus offering a robust platform for evaluating MRC models. The Retro-Reader achieved new state-of-the-art results on these datasets, outperforming strong baselines like BERT, ALBERT, and ELECTRA.

Significant testing demonstrated that the Retro-Reader's improvements over existing models are statistically significant. The authors employed McNemar’s test to substantiate the advancements in exact match (EM) and F1 score metrics, marking a rigorous evaluation approach not frequently applied to MRC tasks.

Implications and Future Work

The Retro-Reader exemplifies how integrating verification mechanisms tailored to MRC specifics can substantially enhance performance, even when leveraging powerful PrLMs as encoders. This finding challenges the dominant trend of focusing primarily on encoder strength, highlighting the importance of thoughtful decoder architecture and problem-specific strategies.

Looking forward, the implications of the Retro-Reader are notable both practically in terms of application to real-world information retrieval and theoretically in encouraging further exploration into sophisticated verifier designs. The paper hints at potential future directions, including more comprehensive decoder mechanisms that resonate with the impressive utility of advanced PrLMs. This underscores a broader trend in AI research where collaborative improvements across model components can yield superior, application-specific outcomes.

In summary, this research contributes a thoughtfully designed model that elevates machine reading comprehension performance, particularly in discerning the answerability of questions—a critical function that aligns MRC more closely with human-like reading comprehension capabilities.