Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Consolidation via Policy Information Regularization in Deep RL for Multi-Agent Games (2011.11517v1)

Published 23 Nov 2020 in cs.AI

Abstract: This paper introduces an information-theoretic constraint on learned policy complexity in the Multi-Agent Deep Deterministic Policy Gradient (MADDPG) reinforcement learning algorithm. Previous research with a related approach in continuous control experiments suggests that this method favors learning policies that are more robust to changing environment dynamics. The multi-agent game setting naturally requires this type of robustness, as other agents' policies change throughout learning, introducing a nonstationary environment. For this reason, recent methods in continual learning are compared to our approach, termed Capacity-Limited MADDPG. Results from experimentation in multi-agent cooperative and competitive tasks demonstrate that the capacity-limited approach is a good candidate for improving learning performance in these environments.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Tyler Malloy (7 papers)
  2. Tim Klinger (23 papers)
  3. Miao Liu (98 papers)
  4. Matthew Riemer (32 papers)
  5. Gerald Tesauro (29 papers)
  6. Chris R. Sims (3 papers)

Summary

We haven't generated a summary for this paper yet.