Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Decentralized Multi-Agents by Imitation of a Centralized Controller (1902.02311v4)

Published 6 Feb 2019 in cs.MA, cs.AI, cs.LG, and cs.SY

Abstract: We consider a multi-agent reinforcement learning problem where each agent seeks to maximize a shared reward while interacting with other agents, and they may or may not be able to communicate. Typically the agents do not have access to other agent policies and thus each agent is situated in a non-stationary and partially-observable environment. In order to obtain multi-agents that act in a decentralized manner, we introduce a novel algorithm under the popular framework of centralized training, but decentralized execution. This training framework first obtains solutions to a multi-agent problem with a single centralized joint-space learner, which is then used to guide imitation learning for independent decentralized multi-agents. This framework has the flexibility to use any reinforcement learning algorithm to obtain the expert as well as any imitation learning algorithm to obtain the decentralized agents. This is in contrast to other multi-agent learning algorithms that, for example, can require more specific structures. We present some theoretical bounds for our method, and we show that one can obtain decentralized solutions to a multi-agent problem through imitation learning.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Alex Tong Lin (10 papers)
  2. Mark J. Debord (2 papers)
  3. Katia Estabridis (2 papers)
  4. Gary Hewer (3 papers)
  5. Guido Montufar (79 papers)
  6. Stanley Osher (104 papers)
Citations (6)

Summary

We haven't generated a summary for this paper yet.