Multilingual Fine-Grained News Headline Hallucination Detection (2407.15975v1)
Abstract: The popularity of automated news headline generation has surged with advancements in pre-trained LLMs. However, these models often suffer from the ``hallucination'' problem, where the generated headline is not fully supported by its source article. Efforts to address this issue have predominantly focused on English, using over-simplistic classification schemes that overlook nuanced hallucination types. In this study, we introduce the first multilingual, fine-grained news headline hallucination detection dataset that contains over 11 thousand pairs in 5 languages, each annotated with detailed hallucination types by experts. We conduct extensive experiments on this dataset under two settings. First, we implement several supervised fine-tuning approaches as preparatory solutions and demonstrate this dataset's challenges and utilities. Second, we test various LLMs' in-context learning abilities and propose two novel techniques, language-dependent demonstration selection and coarse-to-fine prompting, to boost the few-shot hallucination detection performance in terms of the example-F1 metric. We release this dataset to foster further research in multilingual, fine-grained headline hallucination detection.
- Jiaming Shen (56 papers)
- Tianqi Liu (49 papers)
- Jialu Liu (21 papers)
- Zhen Qin (105 papers)
- Jay Pavagadhi (3 papers)
- Simon Baumgartner (10 papers)
- Michael Bendersky (63 papers)