Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
51 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Static network structure cannot stabilize cooperation among Large Language Model agents (2411.10294v1)

Published 15 Nov 2024 in cs.SI, cs.CY, cs.GT, and physics.soc-ph

Abstract: LLMs are increasingly used to model human social behavior, with recent research exploring their ability to simulate social dynamics. Here, we test whether LLMs mirror human behavior in social dilemmas, where individual and collective interests conflict. Humans generally cooperate more than expected in laboratory settings, showing less cooperation in well-mixed populations but more in fixed networks. In contrast, LLMs tend to exhibit greater cooperation in well-mixed settings. This raises a key question: Are LLMs about to emulate human behavior in cooperative dilemmas on networks? In this study, we examine networked interactions where agents repeatedly engage in the Prisoner's Dilemma within both well-mixed and structured network configurations, aiming to identify parallels in cooperative behavior between LLMs and humans. Our findings indicate critical distinctions: while humans tend to cooperate more within structured networks, LLMs display increased cooperation mainly in well-mixed environments, with limited adjustment to networked contexts. Notably, LLM cooperation also varies across model types, illustrating the complexities of replicating human-like social adaptability in artificial agents. These results highlight a crucial gap: LLMs struggle to emulate the nuanced, adaptive social strategies humans deploy in fixed networks. Unlike human participants, LLMs do not alter their cooperative behavior in response to network structures or evolving social contexts, missing the reciprocity norms that humans adaptively employ. This limitation points to a fundamental need in future LLM design -- to integrate a deeper comprehension of social norms, enabling more authentic modeling of human-like cooperation and adaptability in networked environments.

Insights into Cooperative Dynamics in LLM Agents

The paper "Static network structure cannot stabilize cooperation among LLM agents" presents an intriguing investigation into the cooperative behaviors of LLMs when placed in the context of networked social dilemmas, specifically utilizing the Prisoner's Dilemma as a framework. This paper probes whether LLMs can effectively emulate human-like social behavior, particularly in structured networks where the intricacies of network topology and interaction patterns play a critical role.

The authors focus on contrasting well-mixed populations with structured network environments to analyze how LLMs such as GPT-3.5 and GPT-4 manage cooperation. In human-centric studies, individuals typically exhibit more cooperation in structured settings due to repeated interactions fostering trust and cooperative norms. Interestingly, the findings indicate that LLMs deviate from this human behavior pattern. GPT-4, for example, shows a significant inclination towards cooperation in well-mixed, random interaction settings—a surprising result considering that such environments generally promote defection due to lack of predictability in interactions.

A critical aspect of the paper involves examining cooperation dynamics across different network densities, defined by the parameter kk, and varying benefit-to-cost (b/c) ratios. Humans generally adapt their strategies based on these parameters, cooperating more when b/c>kb/c > k. However, GPT-3.5 demonstrates consistent cooperation levels across all scenarios, suggesting a lack of sensitivity to these changes. Contrarily, GPT-4 adapts to some extent, increasing cooperation when advantageous (high b/cb/c) but failing to sustain it as networks become denser.

In exploring these dynamics, the researchers highlight the distinct payoff structures impacting cooperative decisions. For LLMs, defectors consistently outperform cooperators, yet the models' composition remains skewed towards cooperation, particularly in GPT-4, indicating a fundamental divergence from human strategies which often balance based on reciprocal benefits.

Experimental setups where neighbor behaviors are controlled reveal further insights into LLMs' adaptability hazards. While Claude and GPT-4 initially favor cooperation in fully cooperative settings, they quickly shift to defection as neighborhood interactions turn less favorable. This response indicates a degree of adaptability yet falls short of the nuanced strategic adjustments humans would typically make.

The implications of these findings are manifold. They underscore a fundamental gap between human and LLM behavioral models in networked social dilemmas, suggesting limitations in current LLM capabilities to discern and adapt based on complex social cues intrinsic to network interactions. The rigidity in behavior observed among LLMs, particularly their tendency towards period-two cycles, underlines the challenges involved in embedding human-like adaptability and strategic flexibility within AI systems.

This paper advocates for advancing LLM design to integrate a more comprehensive understanding of social norms and network reciprocity, potentially employing backstories or demographic enrichments to align AI behaviors closer to human strategies. Ultimately, while LLMs showcase potential in mirroring specific human-like cooperative strategies in simple settings, their application in complex, networked environments remains severely limited. Future developments might benefit from leveraging hybrid models—combining human intuitive understanding with LLM computational capacity—to bridge existing gaps and harness AI potential more effectively in social and behavioral sciences.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Jin Han (12 papers)
  2. Balaraju Battu (3 papers)
  3. Talal Rahwan (46 papers)
  4. Petter Holme (101 papers)
  5. Ivan Romić (1 paper)