Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

On the Risk of Misinformation Pollution with Large Language Models (2305.13661v2)

Published 23 May 2023 in cs.CL and cs.AI

Abstract: In this paper, we comprehensively investigate the potential misuse of modern LLMs for generating credible-sounding misinformation and its subsequent impact on information-intensive applications, particularly Open-Domain Question Answering (ODQA) systems. We establish a threat model and simulate potential misuse scenarios, both unintentional and intentional, to assess the extent to which LLMs can be utilized to produce misinformation. Our study reveals that LLMs can act as effective misinformation generators, leading to a significant degradation in the performance of ODQA systems. To mitigate the harm caused by LLM-generated misinformation, we explore three defense strategies: prompting, misinformation detection, and majority voting. While initial results show promising trends for these defensive strategies, much more work needs to be done to address the challenge of misinformation pollution. Our work highlights the need for further research and interdisciplinary collaboration to address LLM-generated misinformation and to promote responsible use of LLMs.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Yikang Pan (3 papers)
  2. Liangming Pan (59 papers)
  3. Wenhu Chen (134 papers)
  4. Preslav Nakov (253 papers)
  5. Min-Yen Kan (92 papers)
  6. William Yang Wang (254 papers)
Citations (89)