Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

PMIC: Improving Multi-Agent Reinforcement Learning with Progressive Mutual Information Collaboration (2203.08553v4)

Published 16 Mar 2022 in cs.MA and cs.AI

Abstract: Learning to collaborate is critical in Multi-Agent Reinforcement Learning (MARL). Previous works promote collaboration by maximizing the correlation of agents' behaviors, which is typically characterized by Mutual Information (MI) in different forms. However, we reveal sub-optimal collaborative behaviors also emerge with strong correlations, and simply maximizing the MI can, surprisingly, hinder the learning towards better collaboration. To address this issue, we propose a novel MARL framework, called Progressive Mutual Information Collaboration (PMIC), for more effective MI-driven collaboration. PMIC uses a new collaboration criterion measured by the MI between global states and joint actions. Based on this criterion, the key idea of PMIC is maximizing the MI associated with superior collaborative behaviors and minimizing the MI associated with inferior ones. The two MI objectives play complementary roles by facilitating better collaborations while avoiding falling into sub-optimal ones. Experiments on a wide range of MARL benchmarks show the superior performance of PMIC compared with other algorithms.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (11)
  1. Pengyi Li (9 papers)
  2. Hongyao Tang (28 papers)
  3. Tianpei Yang (25 papers)
  4. Xiaotian Hao (11 papers)
  5. Tong Sang (3 papers)
  6. Yan Zheng (102 papers)
  7. Jianye Hao (185 papers)
  8. Matthew E. Taylor (69 papers)
  9. Wenyuan Tao (3 papers)
  10. Zhen Wang (571 papers)
  11. Fazl Barez (42 papers)
Citations (24)