Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Cooperative Multi-Agent Transfer Learning with Level-Adaptive Credit Assignment (2106.00517v3)

Published 1 Jun 2021 in cs.AI

Abstract: Extending transfer learning to cooperative multi-agent reinforcement learning (MARL) has recently received much attention. In contrast to the single-agent setting, the coordination indispensable in cooperative MARL constrains each agent's policy. However, existing transfer methods focus exclusively on agent policy and ignores coordination knowledge. We propose a new architecture that realizes robust coordination knowledge transfer through appropriate decomposition of the overall coordination into several coordination patterns. We use a novel mixing network named level-adaptive QTransformer (LA-QTransformer) to realize agent coordination that considers credit assignment, with appropriate coordination patterns for different agents realized by a novel level-adaptive Transformer (LA-Transformer) dedicated to the transfer of coordination knowledge. In addition, we use a novel agent network named Population Invariant agent with Transformer (PIT) to realize the coordination transfer in more varieties of scenarios. Extensive experiments in StarCraft II micro-management show that LA-QTransformer together with PIT achieves superior performance compared with state-of-the-art baselines.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (13)
  1. Tianze Zhou (5 papers)
  2. Fubiao Zhang (2 papers)
  3. Kun Shao (29 papers)
  4. Kai Li (313 papers)
  5. Wenhan Huang (8 papers)
  6. Jun Luo (193 papers)
  7. Weixun Wang (31 papers)
  8. Yaodong Yang (169 papers)
  9. Hangyu Mao (37 papers)
  10. Bin Wang (750 papers)
  11. Dong Li (429 papers)
  12. Wulong Liu (38 papers)
  13. Jianye Hao (185 papers)
Citations (14)

Summary

We haven't generated a summary for this paper yet.