Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
149 tokens/sec
GPT-4o
9 tokens/sec
Gemini 2.5 Pro Pro
47 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Strategic Intelligence in Large Language Models: Evidence from evolutionary Game Theory (2507.02618v1)

Published 3 Jul 2025 in cs.AI, cs.CL, and cs.GT

Abstract: Are LLMs a new form of strategic intelligence, able to reason about goals in competitive settings? We present compelling supporting evidence. The Iterated Prisoner's Dilemma (IPD) has long served as a model for studying decision-making. We conduct the first ever series of evolutionary IPD tournaments, pitting canonical strategies (e.g., Tit-for-Tat, Grim Trigger) against agents from the leading frontier AI companies OpenAI, Google, and Anthropic. By varying the termination probability in each tournament (the "shadow of the future"), we introduce complexity and chance, confounding memorisation. Our results show that LLMs are highly competitive, consistently surviving and sometimes even proliferating in these complex ecosystems. Furthermore, they exhibit distinctive and persistent "strategic fingerprints": Google's Gemini models proved strategically ruthless, exploiting cooperative opponents and retaliating against defectors, while OpenAI's models remained highly cooperative, a trait that proved catastrophic in hostile environments. Anthropic's Claude emerged as the most forgiving reciprocator, showing remarkable willingness to restore cooperation even after being exploited or successfully defecting. Analysis of nearly 32,000 prose rationales provided by the models reveals that they actively reason about both the time horizon and their opponent's likely strategy, and we demonstrate that this reasoning is instrumental to their decisions. This work connects classic game theory with machine psychology, offering a rich and granular view of algorithmic decision-making under uncertainty.

Summary

  • The paper demonstrates that LLMs exhibit genuine strategic reasoning in evolutionary IPD tournaments by competing against classic game-theoretic strategies.
  • It employs a multi-phase evolutionary framework to benchmark both basic and advanced models, revealing distinctive strategic fingerprints and adaptation levels.
  • The analysis of move rationales confirms that strategic deliberation drives LLM decision-making, highlighting implications for AI safety and multi-agent systems.

Strategic Intelligence in LLMs: Evidence from Evolutionary Game Theory

This paper presents a comprehensive empirical investigation into the strategic reasoning capabilities of LLMs by embedding them as agents in evolutionary Iterated Prisoner's Dilemma (IPD) tournaments. The paper systematically benchmarks LLMs from OpenAI, Google, and Anthropic against canonical game-theoretic strategies, under a variety of environmental conditions, and analyzes both their quantitative performance and the qualitative content of their move rationales. The work provides a granular view of LLM decision-making under uncertainty, with implications for both AI safety and the emerging field of machine psychology.

Experimental Design and Methodology

The core experimental framework is a series of evolutionary IPD tournaments, each consisting of five phases. In each phase, a population of agents—including both classic strategies (e.g., Tit-for-Tat, Grim Trigger, Bayesian, Random) and LLM-driven agents—engages in round-robin matches. After each phase, agents reproduce in proportion to their average per-move score, with selection pressure amplified by squaring the relative fitness. This evolutionary update rule ensures that successful strategies proliferate, while unsuccessful ones are eliminated.

Key experimental factors include:

  • Model Capability: Both "basic" (e.g., gpt-3.5-turbo, gemini-1.5-flash) and "advanced" (e.g., gpt-4o-mini, gemini-2.5-flash, claude-3-haiku) LLMs are tested.
  • Shadow of the Future: The per-round termination probability is varied (10%, 25%, 75%), directly manipulating the expected match length and thus the incentive structure for cooperation versus defection.
  • Stress Tests: Additional runs introduce persistent mutation (random agent injection) and a "LLM showdown" with only advanced LLMs and a Bayesian agent.

Each LLM agent receives a standardized prompt per move, including the payoff matrix, termination probability, full match history, and instructions to provide a rationale followed by a move (C or D). This design isolates strategic reasoning from parameter inference and enables detailed analysis of the models' internal deliberations.

Quantitative Results

Evolutionary Competitiveness

LLMs consistently survive and often proliferate in the evolutionary tournaments, rarely being eliminated except under extreme conditions (notably, OpenAI's model in the 75% termination regime). Advanced models outperform their basic counterparts, both in average score per move and in evolutionary survival, confirming the impact of model scaling on strategic competence.

Strategic Fingerprints

Distinctive and persistent "strategic fingerprints" are observed for each LLM family:

  • Google's Gemini: Exhibits a ruthlessly strategic style, readily exploiting over-cooperators and retaliating against defectors. Gemini adapts its cooperation rate sharply in response to environmental parameters, defecting almost unconditionally when the shadow of the future is short.
  • OpenAI's Models: Display a highly cooperative, forgiving style, maintaining high cooperation rates even when environmental incentives favor defection. This rigidity leads to catastrophic outcomes in hostile environments (e.g., high termination probability).
  • Anthropic's Claude: Emerges as the most forgiving reciprocator, frequently restoring cooperation even after being exploited, and outperforming OpenAI in head-to-head competition.

These fingerprints are quantified via conditional cooperation probabilities (P(C|CC), P(C|CD), P(C|DC), P(C|DD)), revealing systematic differences in how each model responds to prior outcomes.

Adaptivity and Environmental Sensitivity

Gemini demonstrates marked adaptivity, varying its cooperation rate from over 85% in long-horizon environments to near-zero in one-shot-like conditions. OpenAI, by contrast, remains persistently cooperative across all regimes, failing to adapt to increased incentives for defection. Anthropic's model, while highly cooperative, shows some flexibility and outperforms OpenAI in mixed-agent environments.

Population Dynamics and Stability

The agent ecosystem exhibits a U-shaped stability distribution: maximal stability at intermediate termination probabilities (25%), instability at both low (10%) and high (75%) extremes. Persistent mutation (random agent injection) increases diversity but does not destabilize the population as much as low termination probability. The 75% regime leads to rapid population collapse, with defectors dominating.

Qualitative Analysis of LLM Rationales

A 10% sample of nearly 32,000 LLM-generated rationales was coded for references to time horizon and opponent modeling. Key findings include:

  • Horizon Awareness: Gemini explicitly references the time horizon in over 90% of its rationales in short-horizon environments, and adapts its strategy accordingly. OpenAI mentions the horizon less frequently and, even when it does, often fails to adjust its behavior.
  • Opponent Modeling: Both models engage in opponent modeling in the majority of cases, but Gemini's modeling is more explicit and typological, while OpenAI's is more reactive.
  • Instrumentality of Reasoning: The content of the rationales is tightly coupled to the models' actions, supporting the claim that reasoning is integral to decision-making rather than post-hoc justification.

Notably, Gemini's cooperation rate drops precipitously when it attends to a short time horizon, while OpenAI's remains high regardless of horizon awareness. Both models reduce cooperation when explicitly modeling their opponent, but OpenAI's default is unconditional cooperation in the absence of such modeling.

Theoretical and Practical Implications

Reasoning vs. Memorization

The paper provides strong evidence that LLMs engage in genuine strategic reasoning rather than mere memorization. This is supported by:

  • The novelty and complexity of the tournament environments, which preclude reliance on memorized solutions.
  • The systematic adaptation of strategies to environmental parameters, especially by Gemini.
  • The tight coupling between rationale content and move selection, including occasional reasoning errors that directly influence actions.

Machine Psychology and AI Safety

The observed diversity in strategic fingerprints among LLMs from different vendors has significant implications for multi-agent AI systems and AI safety. The fact that models with similar training data and architectures can develop divergent, persistent strategic styles suggests that fine-tuning and model design choices can have substantial downstream effects on agent behavior in competitive and cooperative settings.

Future Directions

The findings open several avenues for further research:

  • Scaling Laws for Strategic Reasoning: As model size increases, does strategic competence continue to improve, and do models converge towards optimal game-theoretic behavior?
  • Generalization to Other Games: Extending the analysis to other classes of games (e.g., coordination, bargaining, signaling) could reveal the breadth of LLM strategic intelligence.
  • Human-AI Interaction: Investigating how LLMs interact with human agents in similar settings may inform the design of AI systems for negotiation, diplomacy, and other strategic domains.
  • Robustness and Safety: Understanding and controlling the strategic fingerprints of LLMs is critical for deploying them in real-world multi-agent environments, where misaligned incentives or excessive rigidity could have adverse consequences.

Conclusion

This work demonstrates that LLMs are capable, adaptive strategic actors in complex, uncertain environments, with reasoning processes that are both observable and instrumental. The diversity of strategic fingerprints across models underscores the importance of careful evaluation and selection of AI agents for deployment in multi-agent systems. The integration of evolutionary game theory with large-scale LLMing provides a powerful framework for probing the emergent properties of artificial strategic intelligence, with broad implications for AI research and application.

Youtube Logo Streamline Icon: https://streamlinehq.com