Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Correcting Diverse Factual Errors in Abstractive Summarization via Post-Editing and Language Model Infilling (2210.12378v2)

Published 22 Oct 2022 in cs.CL

Abstract: Abstractive summarization models often generate inconsistent summaries containing factual errors or hallucinated content. Recent works focus on correcting factual errors in generated summaries via post-editing. Such correction models are trained using adversarial non-factual summaries constructed using heuristic rules for injecting errors. However, generating non-factual summaries using heuristics often does not generalize well to actual model errors. In this work, we propose to generate hard, representative synthetic examples of non-factual summaries through infilling LLMs. With this data, we train a more robust fact-correction model to post-edit the summaries to improve factual consistency. Through quantitative and qualitative experiments on two popular summarization datasets -- CNN/DM and XSum -- we show that our approach vastly outperforms prior methods in correcting erroneous summaries. Our model -- FactEdit -- improves factuality scores by over ~11 points on CNN/DM and over ~31 points on XSum on average across multiple summarization models, producing more factual summaries while maintaining competitive summarization quality.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Vidhisha Balachandran (31 papers)
  2. Hannaneh Hajishirzi (176 papers)
  3. William W. Cohen (79 papers)
  4. Yulia Tsvetkov (142 papers)
Citations (42)
Github Logo Streamline Icon: https://streamlinehq.com

GitHub