Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Continual Domain Adaptation for Machine Reading Comprehension (2008.10874v1)

Published 25 Aug 2020 in cs.IR

Abstract: Machine reading comprehension (MRC) has become a core component in a variety of NLP applications such as question answering and dialogue systems. It becomes a practical challenge that an MRC model needs to learn in non-stationary environments, in which the underlying data distribution changes over time. A typical scenario is the domain drift, i.e. different domains of data come one after another, where the MRC model is required to adapt to the new domain while maintaining previously learned ability. To tackle such a challenge, in this work, we introduce the \textit{Continual Domain Adaptation} (CDA) task for MRC. So far as we know, this is the first study on the continual learning perspective of MRC. We build two benchmark datasets for the CDA task, by re-organizing existing MRC collections into different domains with respect to context type and question type, respectively. We then analyze and observe the catastrophic forgetting (CF) phenomenon of MRC under the CDA setting. To tackle the CDA task, we propose several BERT-based continual learning MRC models using either regularization-based methodology or dynamic-architecture paradigm. We analyze the performance of different continual learning MRC models under the CDA task and show that the proposed dynamic-architecture based model achieves the best performance.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Lixin Su (15 papers)
  2. Jiafeng Guo (161 papers)
  3. Ruqing Zhang (60 papers)
  4. Yixing Fan (55 papers)
  5. Yanyan Lan (87 papers)
  6. Xueqi Cheng (274 papers)
Citations (9)

Summary

We haven't generated a summary for this paper yet.