Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Remember and Forget Experience Replay for Multi-Agent Reinforcement Learning (2203.13319v3)

Published 24 Mar 2022 in cs.LG, cs.AI, math.OC, and stat.ML

Abstract: We present the extension of the Remember and Forget for Experience Replay (ReF-ER) algorithm to Multi-Agent Reinforcement Learning (MARL). ReF-ER was shown to outperform state of the art algorithms for continuous control in problems ranging from the OpenAI Gym to complex fluid flows. In MARL, the dependencies between the agents are included in the state-value estimator and the environment dynamics are modeled via the importance weights used by ReF-ER. In collaborative environments, we find the best performance when the value is estimated using individual rewards and we ignore the effects of other actions on the transition map. We benchmark the performance of ReF-ER MARL on the Stanford Intelligent Systems Laboratory (SISL) environments. We find that employing a single feed-forward neural network for the policy and the value function in ReF-ER MARL, outperforms state of the art algorithms that rely on complex neural network architectures.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Pascal Weber (11 papers)
  2. Daniel Wälchli (3 papers)
  3. Mustafa Zeqiri (3 papers)
  4. Petros Koumoutsakos (68 papers)
Citations (5)

Summary

We haven't generated a summary for this paper yet.