PcLast: Discovering Plannable Continuous Latent States (2311.03534v2)
Abstract: Goal-conditioned planning benefits from learned low-dimensional representations of rich observations. While compact latent representations typically learned from variational autoencoders or inverse dynamics enable goal-conditioned decision making, they ignore state reachability, hampering their performance. In this paper, we learn a representation that associates reachable states together for effective planning and goal-conditioned policy learning. We first learn a latent representation with multi-step inverse dynamics (to remove distracting information), and then transform this representation to associate reachable states together in $\ell_2$ space. Our proposals are rigorously tested in various simulation testbeds. Numerical results in reward-based settings show significant improvements in sampling efficiency. Further, in reward-free settings this approach yields layered state abstractions that enable computationally efficient hierarchical planning for reaching ad hoc goals with zero additional samples.
- Anurag Koul (6 papers)
- Shivakanth Sujit (5 papers)
- Shaoru Chen (18 papers)
- Ben Evans (10 papers)
- Lili Wu (11 papers)
- Byron Xu (2 papers)
- Rajan Chari (4 papers)
- Riashat Islam (30 papers)
- Raihan Seraj (8 papers)
- Yonathan Efroni (38 papers)
- Lekan Molu (12 papers)
- John Langford (94 papers)
- Alex Lamb (45 papers)
- Miro Dudik (1 paper)