Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

PTDE: Personalized Training with Distilled Execution for Multi-Agent Reinforcement Learning (2210.08872v2)

Published 17 Oct 2022 in cs.AI, cs.LG, and cs.MA

Abstract: Centralized Training with Decentralized Execution (CTDE) has emerged as a widely adopted paradigm in multi-agent reinforcement learning, emphasizing the utilization of global information for learning an enhanced joint $Q$-function or centralized critic. In contrast, our investigation delves into harnessing global information to directly enhance individual $Q$-functions or individual actors. Notably, we discover that applying identical global information universally across all agents proves insufficient for optimal performance. Consequently, we advocate for the customization of global information tailored to each agent, creating agent-personalized global information to bolster overall performance. Furthermore, we introduce a novel paradigm named Personalized Training with Distilled Execution (PTDE), wherein agent-personalized global information is distilled into the agent's local information. This distilled information is then utilized during decentralized execution, resulting in minimal performance degradation. PTDE can be seamlessly integrated with state-of-the-art algorithms, leading to notable performance enhancements across diverse benchmarks, including the SMAC benchmark, Google Research Football (GRF) benchmark, and Learning to Rank (LTR) task.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Yiqun Chen (20 papers)
  2. Hangyu Mao (37 papers)
  3. Tianle Zhang (22 papers)
  4. Shiguang Wu (15 papers)
  5. Bin Zhang (227 papers)
  6. Hongxing Chang (1 paper)
  7. Jiaxin Mao (47 papers)
  8. Wei Yang (349 papers)
Citations (7)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets