Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

MASER: Multi-Agent Reinforcement Learning with Subgoals Generated from Experience Replay Buffer (2206.10607v1)

Published 20 Jun 2022 in cs.LG and cs.AI

Abstract: In this paper, we consider cooperative multi-agent reinforcement learning (MARL) with sparse reward. To tackle this problem, we propose a novel method named MASER: MARL with subgoals generated from experience replay buffer. Under the widely-used assumption of centralized training with decentralized execution and consistent Q-value decomposition for MARL, MASER automatically generates proper subgoals for multiple agents from the experience replay buffer by considering both individual Q-value and total Q-value. Then, MASER designs individual intrinsic reward for each agent based on actionable representation relevant to Q-learning so that the agents reach their subgoals while maximizing the joint action value. Numerical results show that MASER significantly outperforms StarCraft II micromanagement benchmark compared to other state-of-the-art MARL algorithms.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Jeewon Jeon (1 paper)
  2. Woojun Kim (20 papers)
  3. Whiyoung Jung (7 papers)
  4. Youngchul Sung (48 papers)
Citations (32)

Summary

We haven't generated a summary for this paper yet.