Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Globally Optimal Hierarchical Reinforcement Learning for Linearly-Solvable Markov Decision Processes (2106.15380v3)

Published 29 Jun 2021 in cs.LG and cs.AI

Abstract: In this work we present a novel approach to hierarchical reinforcement learning for linearly-solvable Markov decision processes. Our approach assumes that the state space is partitioned, and the subtasks consist in moving between the partitions. We represent value functions on several levels of abstraction, and use the compositionality of subtasks to estimate the optimal values of the states in each partition. The policy is implicitly defined on these optimal value estimates, rather than being decomposed among the subtasks. As a consequence, our approach can learn the globally optimal policy, and does not suffer from the non-stationarity of high-level decisions. If several partitions have equivalent dynamics, the subtasks of those partitions can be shared. If the set of boundary states is smaller than the entire state space, our approach can have significantly smaller sample complexity than that of a flat learner, and we validate this empirically in several experiments.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Guillermo Infante (3 papers)
  2. Anders Jonsson (47 papers)
  3. Vicenç Gómez (39 papers)
Citations (5)

Summary

We haven't generated a summary for this paper yet.