Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

From Cloze to Comprehension: Retrofitting Pre-trained Masked Language Model to Pre-trained Machine Reader (2212.04755v3)

Published 9 Dec 2022 in cs.CL

Abstract: We present Pre-trained Machine Reader (PMR), a novel method for retrofitting pre-trained masked LLMs (MLMs) to pre-trained machine reading comprehension (MRC) models without acquiring labeled data. PMR can resolve the discrepancy between model pre-training and downstream fine-tuning of existing MLMs. To build the proposed PMR, we constructed a large volume of general-purpose and high-quality MRC-style training data by using Wikipedia hyperlinks and designed a Wiki Anchor Extraction task to guide the MRC-style pre-training. Apart from its simplicity, PMR effectively solves extraction tasks, such as Extractive Question Answering and Named Entity Recognition. PMR shows tremendous improvements over existing approaches, especially in low-resource scenarios. When applied to the sequence classification task in the MRC formulation, PMR enables the extraction of high-quality rationales to explain the classification process, thereby providing greater prediction explainability. PMR also has the potential to serve as a unified model for tackling various extraction and classification tasks in the MRC formulation.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Weiwen Xu (19 papers)
  2. Xin Li (980 papers)
  3. Wenxuan Zhang (75 papers)
  4. Meng Zhou (33 papers)
  5. Wai Lam (117 papers)
  6. Luo Si (73 papers)
  7. Lidong Bing (144 papers)
Citations (2)

Summary

We haven't generated a summary for this paper yet.