Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Value Function Spaces: Skill-Centric State Abstractions for Long-Horizon Reasoning (2111.03189v2)

Published 4 Nov 2021 in cs.LG, cs.AI, and cs.RO

Abstract: Reinforcement learning can train policies that effectively perform complex tasks. However for long-horizon tasks, the performance of these methods degrades with horizon, often necessitating reasoning over and chaining lower-level skills. Hierarchical reinforcement learning aims to enable this by providing a bank of low-level skills as action abstractions. Hierarchies can further improve on this by abstracting the space states as well. We posit that a suitable state abstraction should depend on the capabilities of the available lower-level policies. We propose Value Function Spaces: a simple approach that produces such a representation by using the value functions corresponding to each lower-level skill. These value functions capture the affordances of the scene, thus forming a representation that compactly abstracts task relevant information and robustly ignores distractors. Empirical evaluations for maze-solving and robotic manipulation tasks demonstrate that our approach improves long-horizon performance and enables better zero-shot generalization than alternative model-free and model-based methods.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Dhruv Shah (48 papers)
  2. Peng Xu (357 papers)
  3. Yao Lu (212 papers)
  4. Ted Xiao (40 papers)
  5. Alexander Toshev (48 papers)
  6. Sergey Levine (531 papers)
  7. Brian Ichter (52 papers)
Citations (39)

Summary

We haven't generated a summary for this paper yet.