Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Bad Actor, Good Advisor: Exploring the Role of Large Language Models in Fake News Detection (2309.12247v2)

Published 21 Sep 2023 in cs.CL, cs.AI, and cs.CY
Bad Actor, Good Advisor: Exploring the Role of Large Language Models in Fake News Detection

Abstract: Detecting fake news requires both a delicate sense of diverse clues and a profound understanding of the real-world background, which remains challenging for detectors based on small LLMs (SLMs) due to their knowledge and capability limitations. Recent advances in LLMs have shown remarkable performance in various tasks, but whether and how LLMs could help with fake news detection remains underexplored. In this paper, we investigate the potential of LLMs in fake news detection. First, we conduct an empirical study and find that a sophisticated LLM such as GPT 3.5 could generally expose fake news and provide desirable multi-perspective rationales but still underperforms the basic SLM, fine-tuned BERT. Our subsequent analysis attributes such a gap to the LLM's inability to select and integrate rationales properly to conclude. Based on these findings, we propose that current LLMs may not substitute fine-tuned SLMs in fake news detection but can be a good advisor for SLMs by providing multi-perspective instructive rationales. To instantiate this proposal, we design an adaptive rationale guidance network for fake news detection (ARG), in which SLMs selectively acquire insights on news analysis from the LLMs' rationales. We further derive a rationale-free version of ARG by distillation, namely ARG-D, which services cost-sensitive scenarios without querying LLMs. Experiments on two real-world datasets demonstrate that ARG and ARG-D outperform three types of baseline methods, including SLM-based, LLM-based, and combinations of small and LLMs.

The Role of LLMs in Enhancing Fake News Detection: An Examination of LLMs as Advisors

LLMs have garnered significant attention for their remarkable proficiency across a diverse array of tasks. However, their specific applicability in domains requiring nuanced understanding, such as fake news detection, remains a topic of exploration. This essay reviews a paper that critically evaluates the role of LLMs in fake news detection and proposes a methodological framework that leverages the strengths of both LLMs and Small LLMs (SLMs).

Main Findings and Methodology

The paper starts by empirically examining whether LLMs like GPT-3.5 can effectively detect fake news. It contrasts the performance of LLMs across various prompting strategies with a fine-tuned SLM, particularly BERT. The investigation concludes that despite LLMs' prowess in reason generation and multi-perspective analysis, they fall short of outperforming SLMs in task-specific roles such as fake news detection.

Interestingly, the analysis attributes this underperformance to the LLMs' intrinsic limitation in synthesizing and utilizing rationales for final judgment effectively. Consequently, the LLMs, while incapable of surpassing SLMs in overall detection capability, can serve as valuable advisors. They provide insightful rationales that aid in the enhancement of the decision-making process of SLMs.

Based on these insights, the authors propose an Adaptive Rationale Guidance Network (ARG) which mediates the interaction between SLMs and LLMs. The ARG makes use of rationale inputs from LLMs to support SLM decisions in fake news detection. Furthermore, the authors introduce a distilled version, ARG-D, which encapsulates the rationale benefits in a model free from continuous LLM queries, thereby reducing operational costs in scenarios where resource efficiency is crucial.

Empirical Evaluation

The performance of the proposed models, ARG and ARG-D, demonstrates a noteworthy improvement over both standalone SLMs and conventional techniques that combine LLMs with SLMs. Evaluations on real-world datasets from Chinese and English sources substantiate these improvements, showcasing ARG's superior macro F1 scores, thereby validating its efficacy in harnessing rationale guidance from LLMs without directly relying on them for final judgments.

Implications and Future Directions

This research raises several implications for the future development of AI in tasks contingent on extensive real-world knowledge and nuanced reasoning. By establishing LLMs as advisors rather than sole decision-makers, there exists a potential paradigm where the complexities of intricate tasks are managed through collaborative model strategies rather than monolithic model dependencies.

For future research, developing advanced models to parse and refine the integration of multi-perspective rationales from LLMs may offer further improvements. Expanding the exploration to other forms of data and languages could also provide comprehensive insights towards universal applicability and efficiency.

Conclusion

The paper presents a compelling argument for leveraging large-scale LLM capabilities not as standalone replacements for domain-specific models but as insightful collaborators that enhance the effectiveness of small-scale, task-optimized models like BERT. By formulating a hybrid approach, it opens pathways for more robust and resource-efficient deployment strategies in AI applications concerning fake news detection and potentially other domains that require high-contextual understanding.

In sum, while LLMs have yet to fully supersede finely-tuned SLMs on task-specific grounds, their role as strategic aides in providing rationale and enhancing model interpretability marks an important development in computational approaches to artificial intelligence.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (50)
  1. Anthropic. 2023. Model card and evaluations for claude models. https://www-files.anthropic.com/production/images/Model-Card-Claude-2.pdf. Accessed: 2023-08-13.
  2. Language models are few-shot learners. In Advances in Neural Information Processing Systems, pages 1877–1901. Curran Associates Inc.
  3. Kevin Matthe Caramancion. 2023. News verifiers showdown: A comparative performance evaluation of ChatGPT 3.5, ChatGPT 4.0, bing AI, and bard in news fact-checking. arXiv preprint arXiv:2306.17176.
  4. CHEQ. 2019. The economic cost of bad actors on the internet. https://info.cheq.ai/hubfs/Research/THE_ECONOMIC_COST_Fake_News_final.pdf. Accessed: 2023-08-13.
  5. Meta-path-based fake news detection leveraging multi-level social context information. In Proceedings of the 31st ACM International Conference on Information & Knowledge Management, pages 325–334. ACM.
  6. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171–4186. ACL.
  7. Pizzagate: From rumor, to hashtag, to gunfire in dc. The Washington Post.
  8. Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531.
  9. Learn over past, evolve for future: Forecasting temporal trends for fake news detection. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 5: Industry Track), pages 116–125. ACL.
  10. Deep learning for fake news detection: A comprehensive survey. AI Open, 3:133–155.
  11. CHEF: A pilot Chinese dataset for evidence-based fact-checking. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 3362–3376. ACL.
  12. Survey of hallucination in natural language generation. ACM Computing Surveys, 55:1–38.
  13. FakeBERT: Fake news detection in social media with a BERT-based deep learning approach. Multimedia tools and applications, 80(8):11765–11788.
  14. Diederik P Kingma and Jimmy Ba. 2014. Adam: A method for stochastic optimization. In International Conference on Learning Representations.
  15. ChatGPT: Jack of all trades, master of none. Information Fusion, 99:101861.
  16. Large language models are zero-shot reasoners. In Advances in Neural Information Processing Systems, volume 35, pages 22199–22213. Curran Associates, Inc.
  17. Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. ACM Computing Surveys, 55(9):1–35.
  18. Jailbreaking ChatGPT via prompt engineering: An empirical study. arXiv preprint arXiv:2305.13860.
  19. RoBERTa: A robustly optimized BERT pretraining approach. arXiv preprint arXiv:1907.11692.
  20. Large language model is not a good few-shot information extractor, but a good reranker for hard samples! arXiv preprint arXiv:2303.08559.
  21. Divide-and-conquer: Post-user interaction network for fake news detection on social media. In Proceedings of the ACM Web Conference 2022, pages 1148–1158. ACM.
  22. Domain adaptive fake news detection via reinforcement learning. In Proceedings of the ACM Web Conference 2022, pages 3632–3640. ACM.
  23. It’s about time: Rethinking evaluation on rumor detection benchmarks using chronological splits. In Findings of the Association for Computational Linguistics: EACL 2023, pages 736–743. ACL.
  24. Salman Bin Naeem and Rubina Bhatti. 2020. The COVID-19 ‘infodemic’: a new front for information professionals. Health Information & Libraries Journal, 37(3):233–239.
  25. MDFEND: Multi-domain fake news detection. In Proceedings of the 30th ACM International Conference on Information and Knowledge Management. ACM.
  26. FANG: Leveraging social context for fake news detection using graph representation. In Proceedings of the 29th ACM International Conference on Information and Knowledge Management, pages 1165–1174. ACM.
  27. OpenAI. 2022. ChatGPT: Optimizing language models for dialogue. https://openai.com/blog/chatgpt/. Accessed: 2023-08-13.
  28. Towards reliable misinformation mitigation: Generalization, uncertainty, and GPT-4. arXiv preprint arXiv:2305.14928v1.
  29. DeClarE: Debunking fake news and false claims using evidence-aware deep learning. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 22–32. ACL.
  30. Piotr Przybyla. 2020. Capturing the style of fake news. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pages 490–497. AAAI Press.
  31. Improving fake news detection by using an entity-enhanced framework to fuse diverse multimodal clues. In Proceedings of the 29th ACM International Conference on Multimedia, pages 1212–1220. ACM.
  32. Sunil Ramlochan. 2023. Role-playing in large language models like ChatGPT. https://www.promptengineering.org/role-playing-in-large-language-models-like-chatgpt/. Accessed: 2023-08-13.
  33. Yoel Roth. 2022. The vast majority of content we take action on for misinformation is identified proactively. https://twitter.com/yoyoel/status/1483094057471524867. Accessed: 2023-08-13.
  34. Zoom out and observe: News environment perception for fake news detection. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 4543–4556. ACL.
  35. Integrating pattern-and fact-based fake news detection via model preference learning. In Proceedings of the 30th ACM international conference on information & knowledge management, pages 1640–1650. ACM.
  36. dEFEND: Explainable fake news detection. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pages 395–405. ACM.
  37. FakeNewsNet: A data repository with news content, social context and spatiotemporal information for studying fake news on social media. Big data, 8:171–188.
  38. Fake news detection on social media: A data mining perspective. ACM SIGKDD Explorations Newsletter, 19:22–36.
  39. LLaMA: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971.
  40. EANN: Event adversarial neural networks for multi-modal fake news detection. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pages 849–857. ACM.
  41. Emergent abilities of large language models. Transactions on Machine Learning Research.
  42. Chain-of-thought prompting elicits reasoning in large language models. In Advances in Neural Information Processing Systems, volume 35, pages 24824–24837. Curran Associates, Inc.
  43. Transformers: State-of-the-art natural language processing. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pages 38–45, Online. ACL.
  44. Small models are valuable plug-ins for large language models. arXiv preprint arXiv:2305.08848.
  45. Mining dual emotion for fake news detection. In Proceedings of the web conference 2021, pages 3465–3476. ACM.
  46. Siren’s song in the AI ocean: A survey on hallucination in large language models. arXiv preprint arXiv:2309.01219.
  47. A survey of large language models. arXiv preprint arXiv:2303.18223.
  48. Can ChatGPT understand too? a comparative study on ChatGPT and fine-tuned BERT. arXiv preprint arXiv:2302.10198.
  49. Xinyi Zhou and Reza Zafarani. 2019. Network-based fake news detection: A pattern-driven approach. ACM SIGKDD Explorations Newsletter, 21(2):48–60.
  50. Generalizing to the future: Mitigating entity bias in fake news detection. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, pages 2120–2125. ACM.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Beizhe Hu (7 papers)
  2. Qiang Sheng (29 papers)
  3. Juan Cao (73 papers)
  4. Yuhui Shi (44 papers)
  5. Yang Li (1140 papers)
  6. Danding Wang (21 papers)
  7. Peng Qi (55 papers)
Citations (49)
Github Logo Streamline Icon: https://streamlinehq.com

GitHub

  1. GitHub - ICTMCG/ARG (100 stars)