LLMs Can Impersonate Politicians and Other Public Figures
The paper "LLMs can impersonate politicians and other public figures" by Steffen Herbold, Alexander Trautsch, Zlata Kikteva, and Annette Hautli-Janisz tackles the potential and risks associated with the use of LLMs in political discourse. The authors rigorously explore the capability of LLMs, particularly ChatGPT, to impersonate public figures and generate political content that passes as authentic to the general public. This paper addresses crucial gaps in the literature by evaluating how the public perceives these generated impersonations in terms of authenticity, relevance, and coherence.
Study Overview
The central objectives of this paper are captured in two research questions: (1) how do citizens rate the authenticity, relevance, and coherence of LLM-generated debate responses compared to actual responses, and (2) what impact does exposure to these AI-generated responses have on the public's attitude towards AI in public debates. The researchers formed a dataset from thirty episodes of BBC's "Question Time," covering responses from a diverse set of public figures, and used this data to assess the LLM's impersonation capabilities.
Methodology
The paper utilized ChatGPT 4 Turbo to generate debate responses by prompting it with biographical information and specific questions previously addressed by the actual figures. A comprehensive survey was administered to a representative sample of British citizens (n=948), who evaluated these responses without knowing that some were AI-generated. The assessment metrics included authenticity, coherence, relevance, and content similarity, alongside the participants' familiarity with the speakers and their confidence in their judgments.
Results
Impersonation Credibility
The results presented a significant finding: LLM-generated responses were frequently rated higher in authenticity, coherence, and relevance compared to actual responses, as depicted in the statistical distributions and effect sizes (authenticity d=0.66, coherence d=1.25, relevance d=1.23). When participants directly compared actual and AI-generated responses, the perceived authenticity differences narrowed (d=0.22), but the coherence and relevance advantages remained robust.
Content and Linguistic Analysis
Interestingly, the authenticity ratings were generally unaffected by content similarity, with many participants unable to distinguish between genuine and AI-generated responses even when the content differed significantly. Linguistic analysis revealed noteworthy stylistic differences: AI responses exhibited higher lexical diversity and nominalization rates, fewer epistemic markers, and a more substantial overlap between question and response vocabulary.
Public Perception
Public opinion on AI's role in public debates shifted post-exposure to AI's capabilities. Initially skeptical about the utility of AI in debates, participants' appreciation for AI increased after recognizing the quality of AI-generated responses. However, concerns about transparency and regulation were prominent, with over 85% of participants advocating for explicit disclosure and comprehensive regulation of AI use in public discourse.
Implications
The implications of these findings are multifaceted. Practically, the ability of LLMs to generate coherent and relevant political discourse at scale poses risks of misinformation and authenticity confusion. Theoretically, the paper challenges the understanding of human-AI interaction in political communication, suggesting the necessity for stronger content moderation mechanisms and regulatory frameworks.
Conclusion and Future Directions
This research underscores the dual-edged nature of LLMs in political discourse. While they represent an advancement in text generation technologies, their potential to undermine trust and authenticity in political communication is concerning. Future research should focus on enhancing detection methods for AI-generated content and developing policies that ensure the responsible use of AI in public spheres. As LLMs continue to evolve, their role in shaping public opinion necessitates careful examination and proactive governance.