Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Self-Play Posterior Sampling Algorithm for Zero-Sum Markov Games (2210.01907v1)

Published 4 Oct 2022 in cs.LG, cs.AI, and cs.GT

Abstract: Existing studies on provably efficient algorithms for Markov games (MGs) almost exclusively build on the "optimism in the face of uncertainty" (OFU) principle. This work focuses on a different approach of posterior sampling, which is celebrated in many bandits and reinforcement learning settings but remains under-explored for MGs. Specifically, for episodic two-player zero-sum MGs, a novel posterior sampling algorithm is developed with general function approximation. Theoretical analysis demonstrates that the posterior sampling algorithm admits a $\sqrt{T}$-regret bound for problems with a low multi-agent decoupling coefficient, which is a new complexity measure for MGs, where $T$ denotes the number of episodes. When specialized to linear MGs, the obtained regret bound matches the state-of-the-art results. To the best of our knowledge, this is the first provably efficient posterior sampling algorithm for MGs with frequentist regret guarantees, which enriches the toolbox for MGs and promotes the broad applicability of posterior sampling.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Wei Xiong (172 papers)
  2. Han Zhong (38 papers)
  3. Chengshuai Shi (18 papers)
  4. Cong Shen (98 papers)
  5. Tong Zhang (569 papers)
Citations (18)

Summary

We haven't generated a summary for this paper yet.