Papers
Topics
Authors
Recent
2000 character limit reached

Understanding Online Polarization Through Human-Agent Interaction in a Synthetic LLM-Based Social Network (2506.15866v1)

Published 18 Jun 2025 in cs.SI

Abstract: The rise of social media has fundamentally transformed how people engage in public discourse and form opinions. While these platforms offer unprecedented opportunities for democratic engagement, they have been implicated in increasing social polarization and the formation of ideological echo chambers. Previous research has primarily relied on observational studies of social media data or theoretical modeling approaches, leaving a significant gap in our understanding of how individuals respond to and are influenced by polarized online environments. Here we present a novel experimental framework for investigating polarization dynamics that allows human users to interact with LLM-based artificial agents in a controlled social network simulation. Through a user study with 122 participants, we demonstrate that this approach can successfully reproduce key characteristics of polarized online discourse while enabling precise manipulation of environmental factors. Our results provide empirical validation of theoretical predictions about online polarization, showing that polarized environments significantly increase perceived emotionality and group identity salience while reducing expressed uncertainty. These findings extend previous observational and theoretical work by providing causal evidence for how specific features of online environments influence user perceptions and behaviors. More broadly, this research introduces a powerful new methodology for studying social media dynamics, offering researchers unprecedented control over experimental conditions while maintaining ecological validity.

Summary

  • The paper demonstrates a novel experimental framework where human interactions with LLM-driven agents uncover causal links between platform features and perceived polarization.
  • It employs a sophisticated agent-based simulation with LLMs to mimic opinion dynamics and assess the impact of recommendation bias on user engagement.
  • Results show that polarized discourse elevates emotional tone, group identity salience, and bias detection while modulating engagement patterns without major opinion shifts.

Understanding Online Polarization Dynamics via Human Interaction with LLM-Driven Synthetic Social Networks

Introduction

The increasing role of social media in public discourse has brought renewed attention to the phenomenon of polarization—the clustering of individuals into ideologically homogeneous groups with limited cross-group interaction. Although both observational studies and theoretical opinion dynamics models have advanced our understanding, major gaps remain: empirical causal evidence connecting specific online platform features to users' perceptions and engagement is lacking due to the infeasibility of controlled experimentation on mainstream platforms and the abstraction of classic simulation models. This paper addresses this limitation by developing and validating a novel experimental framework in which human participants interact with LLM-based artificial agents within a synthetic social network. This setup enables tightly controlled manipulation of environmental factors (polarization, recommendation regimes) while retaining the ecological complexity of open-ended textual interaction.

Methodological Framework

The simulation system combines agent-based modeling (ABM) principles with GPT-4o-mini-driven agents that generate posts, comments, and social interactions. Each agent is parameterized by a continuous opinion value oi[1,1]o_i \in [-1, 1], personality attributes, a profile, and a local interaction history. Both agent generation and messaging employ complex LLM prompting, incorporating explicit opinion values, persona, and social context, to control for degree and expression of polarization. Two regime types are implemented: one with polarized discourse (bimodal oio_i distributions, extreme certainty, affective language, antagonism) and one with moderate discourse (centered oio_i, hedging, inclusive language, uncertainty).

Interactions, including likes, comments, reposts, and follows, are probabilistically modeled to reflect confirmation bias and differential cross-ideological reactivity. The probability function modulates base reactivity by opinion strength and computes a U-shaped response function: agents are most likely to interact when content echoes their stance, with controlled likelihoods of engaged disagreement for more "active" reactions, reflecting both homophilic and heterophilic interaction patterns. Figure 1

Figure 1: Reaction probability functions for different message opinion values omo_m, capturing the likelihood peaks for ideological alignment and controlled cross-ideological engagement.

The synthetic social graph models follows as a directed network built with homophily and influencer amplification, and adaptively updates as participants interact. Edge formation and dissolution reflect selective exposure and network plasticity, while platform recommendations are governed by a controlled mixture of influence-based popularity, explicit recommendation bias (pro-, contra-, or balanced framing), and mild stochasticity.

Experimental Study Protocol

A 2×32 \times 3 factorial user experiment was conducted (N=122, balanced across cells), manipulating (a) agent population polarization (polarized vs. moderate) and (b) recommendation bias (pro-UBI, balanced, contra-UBI) using Universal Basic Income as the focal controversy. Each participant entered an independent instance of the platform—a Twitter/X-like interface prepopulated with LLM agent content and network structure—and freely posted, liked, commented, or followed for 10 minutes. No human-to-human content was present, ensuring precise exposure control. Figure 2

Figure 2: Screenshot of the simulated platform interface showing the Newsfeed, comment, and follow/like mechanics.

Detailed pre/post questionnaires measured shifts in policy opinion, perceived polarization, affective tone, group salience, uncertainty, and recommendation bias, alongside recording granular engagement activities.

Results

Discourse Perception and Psychological Response

The main effects of the polarized discussion environment on user perceptions are strong and robust. Participants in the polarized regime exhibited significantly elevated perceptions of:

  • Polarization (mean difference g=1.36g=1.36)
  • Emotionality (g=1.53g=1.53; ηp2=0.388\eta_p^2 = 0.388)
  • Group Identity Salience (g=0.73g=0.73)
  • Perceived Platform Bias (g=0.69g=0.69)

Conversely, perceived uncertainty—the observation of hedging and epistemic humility—dropped substantially in the polarized condition (g=1.26g=-1.26; ηp2=0.296\eta_p^2=0.296). Figure 3

Figure 3: Interaction plots showing the effects of polarization and recommendation type on dependent variables; blue=polarized, green=unpolarized.

Recommendation bias had a more limited influence except for interaction effects on bias perception and group identity, particularly in pro-bias settings.

Importantly, while these perceptual shifts are statistically and substantively significant, direct measures of opinion change were not similarly affected, aligning with recent evidence that short-term exposure modulates perception and affect more than explicit stance, and that longer or repeated exposures may be required for belief updating.

User Engagement Patterns

Analysis of engagement reveals several notable points:

  • Likes constitute the majority of interactions, followed by follows, reposts, and comments, reflecting typical effort hierarchy in social media.
  • The greatest overall engagement occurred under polarized + contra-recommendation conditions, both in interaction count and in stronger mean opinion shifts (though not reaching significance).
  • Polarized environments reduced commenting, particularly among predominantly moderate participants, but increased passive (liking, following) and more effortful interactions when confronted with challenging recommendations. Figure 4

    Figure 4: Interaction plots of polarization and recommendation type on user engagement. Blue=polarized, green=moderate; note engagement peaks with polarization plus contra-bias.

    Figure 5

    Figure 5: Distribution of interaction types across all polarization and recommendation conditions; likes predominate, with meaningful shifts in follows and comments by regime.

Modeling Observed Dynamics

The paper demonstrates that properly parameterized LLM-based ABMs can generate content that, when presented to humans, elicits perceptual and behavioral responses consistent with established social identity and affective polarization theory [tajfel_integrative_1979, mason_i_2015, iyengar_affect_2012, tornberg_how_2022-2]. The U-shaped probabilistic engagement model successfully mediates cross-ideological interaction asymmetry, while algorithmic recommendation bias exerts conditionally amplified effects.

The findings show that strong cues of group antagonism and affective polarization not only skew perception, but also alter patterns of engagement: user responses to platform polarization manifest in increased bias detection and social categorization, while the combination of "contra" recommendations and polarized style acts as a catalyst for increased engagement, especially in attention-demanding interactions.

Theoretical and Practical Implications

Empirical Causality and Theoretical Validity

This design closes a longstanding gap by providing causal evidence for the impact of platform-level features (environmental polarization, recommendation curation) on individual perception and engagement—something previously only accessible through observational correlations or overly abstracted simulations [chitra_analyzing_2020-1, del_vicario_echo_2015, santos_link_2021-1, de_arruda_modelling_2022-1]. The systematic main effects and interaction patterns mean that theories of group identity, opinion dynamics, and algorithmic curation can be experimentally validated and parameterized based on observed user response distributions, not just inferred from real-world data and competing interpretations.

Synthetic Social Networks for Polarization Research

LLM-driven synthetic agent approaches allow for rigorous, reproducible, and highly configurable testing of platform interventions—e.g., recommendation diversity, exposure control, influencer amplification—complementing empirical field experiments that are impractical or unethical on real-world platforms [chuang_simulating_2024, breum_persuasive_2024, ohagi_polarization_2024, gao_large_2024]. Extending to multiple topics or allowing network structure to evolve in tandem with interactivity and belief updating opens a rich testbed for interface, policy, or moderation experiments.

Broad Implications and Future Directions

  • Platform design: Findings underscore that algorithmic interventions to diversify exposure must simultaneously account for the affective and identity-consolidating nature of polarized discourse; simple content balancing may have counterintuitive effects depending on group identity and emotionality cues.
  • Polarization mechanisms: Results reinforce that perceived polarization and affective cues—rather than policy attitude extremity per se—are primary levers through which platforms affect group cohesion and antagonism. Affective polarization and identity salience, not just positional divergence, are the principal outcomes of polarized online architectures.
  • Methodology: The integration of LLMs into ABMs in a human-in-the-loop context advances the empirical grounding and ecological validity of social simulation, moving beyond "toy" interactions to capture open-ended debate and social signaling. This is a direct extension of recent research on generative agent-based modeling and collective opinion dynamics with LLMs [ghaffarzadegan_generative_2024, chuang_simulating_2024, gao_large_2024, cisneros-velarde_principles_2024].

Limitations and Prospects for AI Research

  • Agent realism: Although human participants reported high (but not universal) realism, certain stylization limitations (response repetitiveness, insufficient diversity) remain a challenge for LLM-driven agents.
  • Temporal and structural generalizability: Effects are demonstrated in short sessions and with a single, moderately familiar issue (UBI); long-term dynamics, topic transfer, and exposure to "harder" entrenched debates must be further evaluated.
  • Network co-evolution: Current implementation holds agent opinions fixed and network structure largely passive; integrating belief updating and dynamic network rewiring would further bridge the gap to real-world polarization and echo chamber formation [sasahara_social_2021-1, baumann_modeling_2020-2, ramaciotti_morales_inferring_2022].
  • Bias and sampling: As in all synthetic environments, agent diversity, demographic realism, and emergent phenomena are constrained by model parameterization and prompting strategies; care must be taken to avoid reinforcing LLM biases or overfitting to artificial discourse styles [qu_performance_2024].

Future research involves extending simulation time scales, enabling opinion updating for both humans and agents, and systematically varying recommendation algorithms and moderation interventions. Integration with recent advances in reinforcement learning, graph-based recommendation, and longitudinal network analysis is also a promising direction [wu_graph_2022, fan_graph_2019-2, zhang_deep_2022].

Conclusion

The presented framework demonstrates that LLM-based synthetic networks can causally induce and measure core features of online polarization, mirroring both affective and group identity phenomena observed in real social platforms. Human participants exposed to polarized, emotionally charged, and group-marked content in LLM-driven microplatforms display robust shifts in perception, bias detection, and selective engagement, validating theoretical models and revealing actionable insights for platform design and intervention. As LLMs continue to advance in text generation, persona management, and social reasoning, their integration into controlled ABMs provides a scalable, reproducible, and ethically robust environment for dissecting the mechanisms and mitigation strategies for harmful online polarization.

Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.