Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Representative Study on Human Detection of Artificially Generated Media Across Countries (2312.05976v1)

Published 10 Dec 2023 in cs.CR, cs.AI, cs.CY, and cs.LG
A Representative Study on Human Detection of Artificially Generated Media Across Countries

Abstract: AI-generated media has become a threat to our digital society as we know it. These forgeries can be created automatically and on a large scale based on publicly available technology. Recognizing this challenge, academics and practitioners have proposed a multitude of automatic detection strategies to detect such artificial media. However, in contrast to these technical advances, the human perception of generated media has not been thoroughly studied yet. In this paper, we aim at closing this research gap. We perform the first comprehensive survey into people's ability to detect generated media, spanning three countries (USA, Germany, and China) with 3,002 participants across audio, image, and text media. Our results indicate that state-of-the-art forgeries are almost indistinguishable from "real" media, with the majority of participants simply guessing when asked to rate them as human- or machine-generated. In addition, AI-generated media receive is voted more human like across all media types and all countries. To further understand which factors influence people's ability to detect generated media, we include personal variables, chosen based on a literature review in the domains of deepfake and fake news research. In a regression analysis, we found that generalized trust, cognitive reflection, and self-reported familiarity with deepfakes significantly influence participant's decision across all media categories.

Introduction

AI-generated media has evolved rapidly, posing significant challenges to digital society by enabling the creation of convincing forgeries. While numerous automated detection techniques have been proposed, the human capacity to discern AI-generated content from authentic media remains under-explored. Addressing this gap, a comprehensive survey was conducted to understand human detection abilities across different countries and media types—specifically audio, image, and text.

Survey and Methodology

A large-scale paper surveyed 3,002 participants from the USA, Germany, and China to assess their ability to differentiate between human- and machine-generated media. Participants encountered an equal mix of real and synthetic media and rated each piece's authenticity. The paper aimed to answer three research questions: Can individuals identify advanced AI-generated media? Do demographic factors affect detection accuracy? And what cognitive factors impact accuracy?

To explore influencing factors on detection capabilities, personal variables such as generalized trust, cognitive reflection, and familiarity with deepfakes were included, drawing from literature on deepfake and fake news research. A regression analysis was performed to determine the impact of these variables on participants' decisions.

Results and Findings

The paper's findings revealed that participants, on average, struggled to correctly identify AI-generated media. Across all media types and countries, accuracy was consistently low, with participants more often than not mistaking synthetic content for human-generated. The average detection rate for images was below 50%, never exceeding 60% for other media types. Additionally, participants generally believed that media presented to them were created by humans, contrary to a 50/50 distribution of real and synthetic samples.

Regression analysis disclosed significant influences of generalized trust, cognitive reflection, and self-reported familiarity with deepfakes across all media categories. Other variables displayed media-dependent impact. This suggests while AI facilities are currently generating highly convincing media, certain personal traits can still sway an individual's detection abilities.

Demographic and Cognitive Influences

Regarding demographic influences, the analysis showed that German participants were notably proficient at distinguishing AI-generated audio content, potentially due to the lower quality of the German synthetic audio samples used. Other demographic variables, such as age and education, had marginal effects on the detection rate.

Cognitive factors played a more definitive role. Trust negatively correlated with the ability to spot machine-generated media while having a positive correlation with identifying human-generated content. System 2 reasoning, measured through the Cognitive Reflection Test, correlated positively with discerning fakes and negatively with recognizing genuine media. Familiarity with deepfakes was found to assist in detecting authentic audio while complicating the identification of fake images and text.

Conclusion

The paper provides a clear depiction of the current challenges faced by humans in distinguishing AI-generated media from genuine content. The findings underscore the need for continued research into automated detection tools and legislative efforts to manage the societal impacts of generated media. The nuanced understanding drawn from the influence of cognitive and demographic factors on detection rates highlights the complex interplay between human psychology and technology.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Joel Frank (5 papers)
  2. Franziska Herbert (6 papers)
  3. Jonas Ricker (8 papers)
  4. Lea Schönherr (23 papers)
  5. Thorsten Eisenhofer (14 papers)
  6. Asja Fischer (63 papers)
  7. Markus Dürmuth (15 papers)
  8. Thorsten Holz (52 papers)
Citations (6)