Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Reinforced Question Rewriting for Conversational Question Answering (2210.15777v2)

Published 27 Oct 2022 in cs.CL and cs.IR

Abstract: Conversational Question Answering (CQA) aims to answer questions contained within dialogues, which are not easily interpretable without context. Developing a model to rewrite conversational questions into self-contained ones is an emerging solution in industry settings as it allows using existing single-turn QA systems to avoid training a CQA model from scratch. Previous work trains rewriting models using human rewrites as supervision. However, such objectives are disconnected with QA models and therefore more human-like rewrites do not guarantee better QA performance. In this paper we propose using QA feedback to supervise the rewriting model with reinforcement learning. Experiments show that our approach can effectively improve QA performance over baselines for both extractive and retrieval QA. Furthermore, human evaluation shows that our method can generate more accurate and detailed rewrites when compared to human annotations.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Zhiyu Chen (60 papers)
  2. Jie Zhao (214 papers)
  3. Anjie Fang (4 papers)
  4. Besnik Fetahu (27 papers)
  5. Oleg Rokhlenko (22 papers)
  6. Shervin Malmasi (40 papers)
Citations (21)