Papers
Topics
Authors
Recent
Search
2000 character limit reached

Population Replacement Conspiracy Theories

Updated 14 January 2026
  • Population Replacement Conspiracy Theories (PRCT) are narratives alleging that native populations are deliberately replaced by migrants through orchestrated plans and provocative metaphors.
  • The dataset comprises 1,617 headlines from Spanish, Italian, and Portuguese media, annotated with binary PRCT labels along with stance and rhetorical biases.
  • Classification experiments using models like mBERT, Llama, and Mistral highlight varied precision, recall, and the challenges of automated detection amid subjective linguistic cues.

Population Replacement Conspiracy Theories (PRCT) designate a category of conspiratorial narratives characterized by the claim that "native populations are being deliberately replaced by migrants." These theories operate by invoking a framing in which migratory processes are not simply demographic or economic, but instead the effect of a coordinated plan—frequently referenced by terms such as “population replacement,” “ethnic substitution,” or metaphorical language (“poisoning our blood”). PRCTs represent a salient component of contemporary hyperpartisan discourse, exhibiting strong correlations with political polarization, institutional distrust, and potential for motivating extremist violence. Their computational detection is integral to the broader objective of misinformation mitigation and is now studied across multiple linguistic and socio-political contexts, notably in PartisanLens, a multilingual benchmark covering Spanish, Italian, and Portuguese (Maggini et al., 7 Jan 2026).

1. Formal Definition and Taxonomy

Within the PartisanLens framework, PRCT is operationalized as a binary label assigned to headlines or statements if they contain any explicit or implicit reference to a malicious, strategic process intended to “replace” an ingroup population with an outgroup (typically migrants). The exact decision rule states: PRCT=True is triggered by the presence of coordinated demographic plot frames or analogous conspiracy metaphors. Examples include both direct phrasing (“Invaders plan ethnic takeover”) and more oblique allusions (“poisoning our blood”).

The annotation taxonomy remains flat (True/False), without subdivision by alleged perpetrator or mechanisms, but is consistently co-annotated with:

  • Hyperpartisan style (binary: one-sided/extreme)
  • Stance (pro / neutral / against)
  • Three span-level rhetorical biases (loaded language, appeal to fear, name-calling)

This delineation ensures consistent cross-lingual operationalization and enables joint analysis with related phenomena such as rhetorical polarization and stance.

2. Dataset Structure and Label Distributions

PartisanLens presents 1,617 news headlines sourced from Spanish (525), Italian (565), and Portuguese (527) media outlets. Each item is manually annotated with PRCT status plus the aforementioned co-occurring labels.

A breakdown of PRCT versus non-PRCT headlines by language is shown below:

Language PRCT=True PRCT=False
SPA 46 479
ITA 160 405
PT 9 518

Across all languages, hyperpartisan narratives are present in 336 (SPA), 337 (ITA), and 367 (PT) headlines. Stance, rhetorical bias, and other label distributions are comprehensively tabulated in the dataset, with loaded-language and appeal-to-fear marking substantial fractions of hyperpartisan content, especially in Spanish and Italian. These figures reveal sharp imbalances in PRCT prevalence by language, a pattern mirrored across stance orientation and rhetorical device distributions.

3. Annotation Protocol and Reliability

Annotation was conducted by nine native speakers (three per language, mixed gender, Master’s+ educational background) in a masked, multi-stage procedure: pilot, refinement, and full annotation rounds, totaling roughly 66 hours. Outlets were masked to avoid bias.

The headline annotation protocol includes:

  • Hyperpartisan: one-sided or highly emotional framing
  • PRCT: presence of demographic replacement conspiratorial narrative per the above definition
  • Stance: pro/neutral/against in reference to immigration policy
  • Span-level rhetorical marking: loaded language, appeal to fear, name-calling

Inter-annotator agreement was quantified using Fleiss' κ:

Language Hyperpartisan PRCT Stance
SPA 0.876 0.880 0.837
ITA 0.706 0.774 0.744
PT 0.770 0.721 0.737

By conventional standards, these κ values denote substantial to almost perfect agreement. The highest consistency is observed in Spanish, especially for PRCT detection, indicating that conspiratorial replacement narratives manifest with salient and cross-annotator-recognizable linguistic features.

4. Baseline Model Benchmarking and Classification Performance

Classification experiments systematically evaluate multilingual transformer architectures, employing both zero/few-shot prompting and supervised fine-tuning:

  • mBERT (fine-tuned, 110M parameters)
  • Llama 3.1–8B (zero-/few-shot; fine-tuned; rationale-augmented)
  • Mistral–NeMo–Base–2407 (12B parameters, identical schemes)
  • Llama 3.3–70B (all four strategies)

Model outputs conform to a standardized JSON format including hyperpartisan, PRCT, and stance keys. Classification quality is measured by macro-averaged precision (P), recall (R), and F1, ameliorating class imbalance effects.

PRCT detection results (excerpted):

Model Precision Recall F1
mBERT (FT) .8589 .8148 .8345
Llama 8B (FS) .7274 .8041 .7540
Llama 70B (ZS) .7814 .8537 .8090
Llama 70B (FS) .8281 .8175 .8226
Nemo (FS) .9325 .7237 .7814

Key insights:

  • PRCT detection outperforms hyperpartisan identification, attributed to a more distinct conspiratorial lexicon.
  • Nemo achieves peak precision but at the cost of recall, indicating a conservative bias.
  • Few-shot prompting with Llama 70B rivals supervised mBERT, while rationale-augmented fine-tuning yields minimal gains (explanations encode decision style rather than conspiracy features).

5. LLM Annotation and Human Alignment

Evaluations of LLMs as automatic annotators involve mean pairwise Cohen’s κ between human–human (H–H), Llama–human (L–H), and gpt-5-mini–human (G–H) pairs:

Label H–H L–H G–H
SPA-PRCT .893 .483 .549
ITA-PRCT .743 .632 .541
PT-PRCT 1.000 .124 .662

LLMs lag behind human agreement, markedly in Spanish and Portuguese. Leading sources of error include:

  • Misinterpretation of sarcasm or irony
  • Inability to resolve polysemy (e.g., “Imigrantes” as both place and demographic category)
  • Cultural rigidity in LLM understanding (e.g., direct mapping of “invadere” to migration contexts)

Persona-based prompting partially mitigates these deficits (see below).

6. Persona-Based Conditioning for Subjectivity Modeling

Each human annotator's demographic and ideology were profiled via Political Compass and PRISM surveys, then rendered into persona prompts for zero-shot LLM classification. Agreement statistics (PRCT κ):

Persona Llama 70B gpt-5-mini
SPA_1 0.5531 0.6252
ITA_1 0.7409 0.7274
PT_1 0.3151 0.6622

Persona conditioning increases PRCT annotation agreement across all languages, particularly for Portuguese with gpt-5-mini (κ up to .662). Stance labeling also improves under persona prompts, though hyperpartisan trends are less consistent. This approach supports a nuanced simulation of individual annotator variability and subjectivity in conspiracy detection.

7. Methodological Implications and Future Directions

Empirical findings demonstrate that PRCT detection is the most tractable among hyperpartisan, stance, and conspiracy-related classification problems, owing to the salience of conspiratorial linguistic motifs. Hyperpartisan style occupies an intermediate tier of difficulty: its distributed lexical markers are frequently interwoven with stance and emotional rhetoric. Stance detection is the most challenging, requiring subtle pragmatic and socio-cultural knowledge.

Contemporary LLMs, including those at 70B parameter scale, remain inferior to human annotators for subtle or contextually-dependent phenomena (e.g., sarcasm, polysemy). Few-shot prompting and rationale-based techniques yield measurable gains, but task-dependent patterns emerge: whereas rationales encode decision structures, they do not directly enhance conspiracy content recognition.

Persona-based model conditioning is an effective vector for modeling subjective dimensions—critical for contexts where misinformation is evaluated through culturally- and ideologically-contingent lenses. A plausible implication is that future architectures should combine stronger context modeling, greater linguistic diversity, and cross-lingual transfer learning to further close the gap between automated and human annotation.

Comprehensive resources, including code, data, and annotation protocols, are archived at https://github.com/MichJoM/PartisanLens, supporting transparent evaluation and extension of PRCT detection in multilingual and cross-cultural settings (Maggini et al., 7 Jan 2026).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Population Replacement Conspiracy Theories (PRCT).