Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Bi-level Actor-Critic for Multi-agent Coordination (1909.03510v3)

Published 8 Sep 2019 in cs.MA

Abstract: Coordination is one of the essential problems in multi-agent systems. Typically multi-agent reinforcement learning (MARL) methods treat agents equally and the goal is to solve the Markov game to an arbitrary Nash equilibrium (NE) when multiple equilibra exist, thus lacking a solution for NE selection. In this paper, we treat agents \emph{unequally} and consider Stackelberg equilibrium as a potentially better convergence point than Nash equilibrium in terms of Pareto superiority, especially in cooperative environments. Under Markov games, we formally define the bi-level reinforcement learning problem in finding Stackelberg equilibrium. We propose a novel bi-level actor-critic learning method that allows agents to have different knowledge base (thus intelligent), while their actions still can be executed simultaneously and distributedly. The convergence proof is given, while the resulting learning algorithm is tested against the state of the arts. We found that the proposed bi-level actor-critic algorithm successfully converged to the Stackelberg equilibria in matrix games and find an asymmetric solution in a highway merge environment.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Haifeng Zhang (59 papers)
  2. Weizhe Chen (20 papers)
  3. Zeren Huang (3 papers)
  4. Minne Li (14 papers)
  5. Yaodong Yang (169 papers)
  6. Weinan Zhang (322 papers)
  7. Jun Wang (992 papers)
Citations (82)

Summary

We haven't generated a summary for this paper yet.