Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Comparison of Self-Play Algorithms Under a Generalized Framework (2006.04471v1)

Published 8 Jun 2020 in cs.AI and cs.GT

Abstract: Throughout scientific history, overarching theoretical frameworks have allowed researchers to grow beyond personal intuitions and culturally biased theories. They allow to verify and replicate existing findings, and to link is connected results. The notion of self-play, albeit often cited in multiagent Reinforcement Learning, has never been grounded in a formal model. We present a formalized framework, with clearly defined assumptions, which encapsulates the meaning of self-play as abstracted from various existing self-play algorithms. This framework is framed as an approximation to a theoretical solution concept for multiagent training. On a simple environment, we qualitatively measure how well a subset of the captured self-play methods approximate this solution when paired with the famous PPO algorithm. We also provide insights on interpreting quantitative metrics of performance for self-play training. Our results indicate that, throughout training, various self-play definitions exhibit cyclic policy evolutions.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Kevin Denamganai (8 papers)
  2. Sam Devlin (32 papers)
  3. Spyridon Samothrakis (9 papers)
  4. James Alfred Walker (13 papers)
  5. Daniel Hernandez (15 papers)
Citations (10)

Summary

We haven't generated a summary for this paper yet.