Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Generating Self-Contained and Summary-Centric Question Answer Pairs via Differentiable Reward Imitation Learning (2109.04689v1)

Published 10 Sep 2021 in cs.CL, cs.AI, and cs.LG

Abstract: Motivated by suggested question generation in conversational news recommendation systems, we propose a model for generating question-answer pairs (QA pairs) with self-contained, summary-centric questions and length-constrained, article-summarizing answers. We begin by collecting a new dataset of news articles with questions as titles and pairing them with summaries of varying length. This dataset is used to learn a QA pair generation model producing summaries as answers that balance brevity with sufficiency jointly with their corresponding questions. We then reinforce the QA pair generation process with a differentiable reward function to mitigate exposure bias, a common problem in natural language generation. Both automatic metrics and human evaluation demonstrate these QA pairs successfully capture the central gists of the articles and achieve high answer accuracy.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Li Zhou (215 papers)
  2. Kevin Small (15 papers)
  3. Yong Zhang (660 papers)
  4. Sandeep Atluri (4 papers)
Citations (2)