Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

HMRL: Hyper-Meta Learning for Sparse Reward Reinforcement Learning Problem (2002.04238v2)

Published 11 Feb 2020 in cs.LG, cs.AI, and stat.ML

Abstract: In spite of the success of existing meta reinforcement learning methods, they still have difficulty in learning a meta policy effectively for RL problems with sparse reward. In this respect, we develop a novel meta reinforcement learning framework called Hyper-Meta RL(HMRL), for sparse reward RL problems. It is consisted with three modules including the cross-environment meta state embedding module which constructs a common meta state space to adapt to different environments; the meta state based environment-specific meta reward shaping which effectively extends the original sparse reward trajectory by cross-environmental knowledge complementarity and as a consequence the meta policy achieves better generalization and efficiency with the shaped meta reward. Experiments with sparse-reward environments show the superiority of HMRL on both transferability and policy learning efficiency.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Yun Hua (11 papers)
  2. Xiangfeng Wang (70 papers)
  3. Bo Jin (57 papers)
  4. Wenhao Li (136 papers)
  5. Junchi Yan (241 papers)
  6. Xiaofeng He (33 papers)
  7. Hongyuan Zha (136 papers)
Citations (6)

Summary

We haven't generated a summary for this paper yet.