Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Machine Reading Comprehension using Case-based Reasoning (2305.14815v4)

Published 24 May 2023 in cs.CL and cs.IR

Abstract: We present an accurate and interpretable method for answer extraction in machine reading comprehension that is reminiscent of case-based reasoning (CBR) from classical AI. Our method (CBR-MRC) builds upon the hypothesis that contextualized answers to similar questions share semantic similarities with each other. Given a test question, CBR-MRC first retrieves a set of similar cases from a nonparametric memory and then predicts an answer by selecting the span in the test context that is most similar to the contextualized representations of answers in the retrieved cases. The semi-parametric nature of our approach allows it to attribute a prediction to the specific set of evidence cases, making it a desirable choice for building reliable and debuggable QA systems. We show that CBR-MRC provides high accuracy comparable with large reader models and outperforms baselines by 11.5 and 8.4 EM on NaturalQuestions and NewsQA, respectively. Further, we demonstrate the ability of CBR-MRC in identifying not just the correct answer tokens but also the span with the most relevant supporting evidence. Lastly, we observe that contexts for certain question types show higher lexical diversity than others and find that CBR-MRC is robust to these variations while performance using fully-parametric methods drops.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Dung Thai (7 papers)
  2. Dhruv Agarwal (17 papers)
  3. Mudit Chaudhary (6 papers)
  4. Wenlong Zhao (18 papers)
  5. Rajarshi Das (27 papers)
  6. Manzil Zaheer (89 papers)
  7. Jay-Yoon Lee (16 papers)
  8. Hannaneh Hajishirzi (176 papers)
  9. Andrew McCallum (132 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.