Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Correction with Backtracking Reduces Hallucination in Summarization (2310.16176v3)

Published 24 Oct 2023 in cs.CL and cs.AI

Abstract: Abstractive summarization aims at generating natural language summaries of a source document that are succinct while preserving the important elements. Despite recent advances, neural text summarization models are known to be susceptible to hallucinating (or more correctly confabulating), that is to produce summaries with details that are not grounded in the source document. In this paper, we introduce a simple yet efficient technique, CoBa, to reduce hallucination in abstractive summarization. The approach is based on two steps: hallucination detection and mitigation. We show that the former can be achieved through measuring simple statistics about conditional word probabilities and distance to context words. Further, we demonstrate that straight-forward backtracking is surprisingly effective at mitigation. We thoroughly evaluate the proposed method with prior art on three benchmark datasets for text summarization. The results show that CoBa is effective and efficient in reducing hallucination, and offers great adaptability and flexibility. Code can be found at https://github.com/zhenzhel/CoBa.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Zhenzhen Liu (10 papers)
  2. Chao Wan (11 papers)
  3. Varsha Kishore (8 papers)
  4. Jin Peng Zhou (28 papers)
  5. Minmin Chen (35 papers)
  6. Kilian Q. Weinberger (105 papers)
Citations (2)
X Twitter Logo Streamline Icon: https://streamlinehq.com