Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

mFACE: Multilingual Summarization with Factual Consistency Evaluation (2212.10622v2)

Published 20 Dec 2022 in cs.CL

Abstract: Abstractive summarization has enjoyed renewed interest in recent years, thanks to pre-trained LLMs and the availability of large-scale datasets. Despite promising results, current models still suffer from generating factually inconsistent summaries, reducing their utility for real-world application. Several recent efforts attempt to address this by devising models that automatically detect factual inconsistencies in machine generated summaries. However, they focus exclusively on English, a language with abundant resources. In this work, we leverage factual consistency evaluation models to improve multilingual summarization. We explore two intuitive approaches to mitigate hallucinations based on the signal provided by a multilingual NLI model, namely data filtering and controlled generation. Experimental results in the 45 languages from the XLSum dataset show gains over strong baselines in both automatic and human evaluation.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Roee Aharoni (35 papers)
  2. Shashi Narayan (35 papers)
  3. Joshua Maynez (28 papers)
  4. Jonathan Herzig (34 papers)
  5. Elizabeth Clark (16 papers)
  6. Mirella Lapata (135 papers)
Citations (36)