Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learning to Share and Hide Intentions using Information Regularization (1808.02093v2)

Published 6 Aug 2018 in cs.AI, cs.IT, cs.LG, cs.MA, math.IT, and stat.ML

Abstract: Learning to cooperate with friends and compete with foes is a key component of multi-agent reinforcement learning. Typically to do so, one requires access to either a model of or interaction with the other agent(s). Here we show how to learn effective strategies for cooperation and competition in an asymmetric information game with no such model or interaction. Our approach is to encourage an agent to reveal or hide their intentions using an information-theoretic regularizer. We consider both the mutual information between goal and action given state, as well as the mutual information between goal and state. We show how to optimize these regularizers in a way that is easy to integrate with policy gradient reinforcement learning. Finally, we demonstrate that cooperative (competitive) policies learned with our approach lead to more (less) reward for a second agent in two simple asymmetric information games.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. DJ Strouse (15 papers)
  2. Max Kleiman-Weiner (20 papers)
  3. Josh Tenenbaum (18 papers)
  4. Matt Botvinick (15 papers)
  5. David Schwab (4 papers)
Citations (57)

Summary

We haven't generated a summary for this paper yet.