Detect, Investigate, Judge and Determine: A Novel LLM-based Framework for Few-shot Fake News Detection (2407.08952v1)
Abstract: Few-Shot Fake News Detection (FS-FND) aims to distinguish inaccurate news from real ones in extremely low-resource scenarios. This task has garnered increased attention due to the widespread dissemination and harmful impact of fake news on social media. LLMs have demonstrated competitive performance with the help of their rich prior knowledge and excellent in-context learning abilities. However, existing methods face significant limitations, such as the Understanding Ambiguity and Information Scarcity, which significantly undermine the potential of LLMs. To address these shortcomings, we propose a Dual-perspective Augmented Fake News Detection (DAFND) model, designed to enhance LLMs from both inside and outside perspectives. Specifically, DAFND first identifies the keywords of each news article through a Detection Module. Subsequently, DAFND creatively designs an Investigation Module to retrieve inside and outside valuable information concerning to the current news, followed by another Judge Module to derive its respective two prediction results. Finally, a Determination Module further integrates these two predictions and derives the final result. Extensive experiments on two publicly available datasets show the efficacy of our proposed method, particularly in low-resource settings.
- AI@Meta. 2024. Llama 3 model card.
- David Boissonneault and Emily Hensen. 2024. Fake news detection with large language models on the liar dataset.
- Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901.
- Glm: General language model pretraining with autoregressive blank infilling. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 320–335.
- Kan: Knowledge-aware attention network for fake news detection. In Proceedings of the AAAI conference on artificial intelligence, volume 35, pages 81–89.
- Making pre-trained language models better few-shot learners. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 3816–3830.
- Fakeflow: Fake news detection by modeling the flow of affective information. In Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, pages 679–689.
- Deberta: Decoding-enhanced bert with disentangled attention. In International Conference on Learning Representations.
- Training compute-optimal large language models. arXiv preprint arXiv:2203.15556.
- Benjamin Horne and Sibel Adali. 2017. This just in: Fake news packs a lot in title, uses simpler, repetitive content in text body, more similar to satire than real news. In Proceedings of the international AAAI conference on web and social media, volume 11, pages 759–766.
- Bad actor, good advisor: Exploring the role of large language models in fake news detection. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 38, pages 22105–22113.
- Lora: Low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685.
- Compare to the knowledge: Graph neural fake news detection with external knowledge. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 754–763.
- Faking fake news for real fake news detection: Propaganda-loaded training data generation. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 14571–14589.
- Fake news detection via knowledgeable prompt learning. Information Processing & Management, 59(5):103029.
- Generalization through memorization: Nearest neighbor language models. In International Conference on Learning Representations.
- What makes good in-context examples for gpt-3? In Proceedings of Deep Learning Inside Out (DeeLIO 2022), pages 100–114.
- Kapalm: Knowledge graph enhanced language models for fake news detection. In Findings of the Association for Computational Linguistics: EMNLP 2023, pages 3999–4009.
- Mdfend: Multi-domain fake news detection. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management, pages 3343–3347.
- Improving generalizability of fake news detection methods using propensity score matching. arXiv preprint arXiv:2002.00838.
- OpenAI. 2023. Gpt-4 technical report.
- Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35:27730–27744.
- Art: Automatic multi-step reasoning and tool-use for large language models. arXiv preprint arXiv:2303.09014.
- Learning to retrieve prompts for in-context learning. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 2655–2671.
- Fakenewsnet: A data repository with news content, social context and spatiotemporal information for studying fake news on social media. Journal on big data, 8(3).
- Fake news detection on social media: A data mining perspective. ACM SIGKDD explorations newsletter, 19(1):22–36.
- Integrating large language models and machine learning for fake news detection. In 2024 20th IEEE International Colloquium on Signal Processing & Its Applications (CSPA), pages 102–107. IEEE.
- Zephyr: Direct distillation of lm alignment. arXiv preprint arXiv:2310.16944.
- Explainable fake news detection with large language model via defense among competing wisdom. In Proceedings of the ACM on Web Conference 2024, pages 2452–2463.
- Transformers: State-of-the-art natural language processing. In Proceedings of the 2020 conference on empirical methods in natural language processing: system demonstrations, pages 38–45.
- Answering questions by meta-reasoning over multiple chains of thought. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 5942–5966.
- Glm-130b: An open bilingual pre-trained model. arXiv preprint arXiv:2210.02414.
- Calibrate before use: Improving few-shot performance of language models. In International Conference on Machine Learning, pages 12697–12706. PMLR.
- Ye Liu (153 papers)
- Jiajun Zhu (16 papers)
- Kai Zhang (542 papers)
- Haoyu Tang (18 papers)
- Yanghai Zhang (7 papers)
- Xukai Liu (5 papers)
- Qi Liu (485 papers)
- Enhong Chen (242 papers)