Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Sample Efficient Reinforcement Learning through Learning from Demonstrations in Minecraft (2003.06066v1)

Published 12 Mar 2020 in cs.LG and stat.ML

Abstract: Sample inefficiency of deep reinforcement learning methods is a major obstacle for their use in real-world applications. In this work, we show how human demonstrations can improve final performance of agents on the Minecraft minigame ObtainDiamond with only 8M frames of environment interaction. We propose a training procedure where policy networks are first trained on human data and later fine-tuned by reinforcement learning. Using a policy exploitation mechanism, experience replay and an additional loss against catastrophic forgetting, our best agent was able to achieve a mean score of 48. Our proposed solution placed 3rd in the NeurIPS MineRL Competition for Sample-Efficient Reinforcement Learning.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Christian Scheller (11 papers)
  2. Yanick Schraner (9 papers)
  3. Manfred Vogel (11 papers)
Citations (25)

Summary

We haven't generated a summary for this paper yet.