Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

MetaCURE: Meta Reinforcement Learning with Empowerment-Driven Exploration (2006.08170v5)

Published 15 Jun 2020 in cs.AI and cs.LG

Abstract: Meta reinforcement learning (meta-RL) extracts knowledge from previous tasks and achieves fast adaptation to new tasks. Despite recent progress, efficient exploration in meta-RL remains a key challenge in sparse-reward tasks, as it requires quickly finding informative task-relevant experiences in both meta-training and adaptation. To address this challenge, we explicitly model an exploration policy learning problem for meta-RL, which is separated from exploitation policy learning, and introduce a novel empowerment-driven exploration objective, which aims to maximize information gain for task identification. We derive a corresponding intrinsic reward and develop a new off-policy meta-RL framework, which efficiently learns separate context-aware exploration and exploitation policies by sharing the knowledge of task inference. Experimental evaluation shows that our meta-RL method significantly outperforms state-of-the-art baselines on various sparse-reward MuJoCo locomotion tasks and more complex sparse-reward Meta-World tasks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Jin Zhang (314 papers)
  2. Jianhao Wang (16 papers)
  3. Hao Hu (114 papers)
  4. Tong Chen (200 papers)
  5. Yingfeng Chen (30 papers)
  6. Changjie Fan (79 papers)
  7. Chongjie Zhang (68 papers)
Citations (25)