Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Meta reinforcement learning as task inference (1905.06424v2)

Published 15 May 2019 in cs.LG, cs.AI, and stat.ML

Abstract: Humans achieve efficient learning by relying on prior knowledge about the structure of naturally occurring tasks. There is considerable interest in designing reinforcement learning (RL) algorithms with similar properties. This includes proposals to learn the learning algorithm itself, an idea also known as meta learning. One formal interpretation of this idea is as a partially observable multi-task RL problem in which task information is hidden from the agent. Such unknown task problems can be reduced to Markov decision processes (MDPs) by augmenting an agent's observations with an estimate of the belief about the task based on past experience. However estimating the belief state is intractable in most partially-observed MDPs. We propose a method that separately learns the policy and the task belief by taking advantage of various kinds of privileged information. Our approach can be very effective at solving standard meta-RL environments, as well as a complex continuous control environment with sparse rewards and requiring long-term memory.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Jan Humplik (15 papers)
  2. Alexandre Galashov (21 papers)
  3. Leonard Hasenclever (33 papers)
  4. Pedro A. Ortega (34 papers)
  5. Yee Whye Teh (162 papers)
  6. Nicolas Heess (139 papers)
Citations (123)

Summary

We haven't generated a summary for this paper yet.