Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

ProQA: Structural Prompt-based Pre-training for Unified Question Answering (2205.04040v2)

Published 9 May 2022 in cs.CL

Abstract: Question Answering (QA) is a longstanding challenge in natural language processing. Existing QA works mostly focus on specific question types, knowledge domains, or reasoning skills. The specialty in QA research hinders systems from modeling commonalities between tasks and generalization for wider applications. To address this issue, we present ProQA, a unified QA paradigm that solves various tasks through a single model. ProQA takes a unified structural prompt as the bridge and improves the QA-centric ability by structural prompt-based pre-training. Through a structurally designed prompt-based input schema, ProQA concurrently models the knowledge generalization for all QA tasks while keeping the knowledge customization for every specific QA task. Furthermore, ProQA is pre-trained with structural prompt-formatted large-scale synthesized corpus, which empowers the model with the commonly-required QA ability. Experimental results on 11 QA benchmarks demonstrate that ProQA consistently boosts performance on both full data fine-tuning, few-shot learning, and zero-shot testing scenarios. Furthermore, ProQA exhibits strong ability in both continual learning and transfer learning by taking the advantages of the structural prompt.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Wanjun Zhong (49 papers)
  2. Yifan Gao (69 papers)
  3. Ning Ding (122 papers)
  4. Yujia Qin (41 papers)
  5. Zhiyuan Liu (433 papers)
  6. Ming Zhou (182 papers)
  7. Jiahai Wang (31 papers)
  8. Jian Yin (67 papers)
  9. Nan Duan (172 papers)
Citations (31)