Learning Planning Abstractions from Language (2405.03864v1)
Abstract: This paper presents a framework for learning state and action abstractions in sequential decision-making domains. Our framework, planning abstraction from language (PARL), utilizes language-annotated demonstrations to automatically discover a symbolic and abstract action space and induce a latent state abstraction based on it. PARL consists of three stages: 1) recovering object-level and action concepts, 2) learning state abstractions, abstract action feasibility, and transition models, and 3) applying low-level policies for abstract actions. During inference, given the task description, PARL first makes abstract action plans using the latent transition and feasibility functions, then refines the high-level plan using low-level policies. PARL generalizes across scenarios involving novel object instances and environments, unseen concept compositions, and tasks that require longer planning horizons than settings it is trained on.
- TAPS: Task-agnostic policy sequencing. In ICRA, 2023.
- Alignment-based compositional semantics for instruction following. In EMNLP, 2015.
- Modular multitask reinforcement learning with policy sketches. In ICML, 2017.
- Hindsight Experience Replay. In NeurIPS, 2017.
- Learning Neural-Symbolic Descriptive Planning Models via Cube-Space Priors: The Voyage Home (To STRIPS). arXiv:2004.12850, 2020.
- Recent advances in hierarchical reinforcement learning. Discrete event dynamic systems, 13(1):41–77, 2003.
- Learning First-Order Symbolic Representations for Planning from the Structure of the State Space. In ECAI, 2020.
- Learning and planning for temporally extended tasks in unknown environments, 2021.
- Decision Transformer: Reinforcement Learning via Sequence Modeling. In NeurIPS, 2021.
- BabyAI: First Steps Towards Grounded Language Learning With a Human In the Loop. In ICLR, 2019.
- Recurrent Environment Simulators. In ICLR, 2017.
- Learning Neuro-Symbolic Relational Transition Models for Bilevel Planning. arXiv:2105.14074, 2021.
- Modular networks for compositional instruction following. In NAACL-HLT, pp. 1033–1040, 2021.
- Pybullet, a python module for physics simulation in robotics, games and machine learning, 2017.
- Tomás de la Rosa and Sheila McIlraith. Learning domain control knowledge for tlplan and beyond. In ICAPS 2011 Workshop on Planning and Learning, 2011.
- Thomas G Dietterich. Hierarchical reinforcement learning with the maxq value function decomposition. JAIR, 13:227–303, 2000.
- Automated Planning: Theory and Practice. Elsevier, 2004.
- A Theory of Abstraction. Artif. Intell., 57(2-3):323–389, 1992.
- Equivalence Notions and Model Minimization in Markov Decision Processes. Artif. Intell., 147(1-2):163–223, 2003.
- Pct: Point cloud transformer. Computational Visual Media, 7:187–199, 2021.
- Latent space planning for multi-object manipulation with environment-aware relational classifiers. arXiv preprint arXiv:2305.10857, 2023.
- Learning Grounded Relational Symbols from Continuous Data for Abstract Reasoning. In ICRA Workshop, 2013.
- Language as an abstraction for hierarchical deep reinforcement learning. Advances in Neural Information Processing Systems, 32, 2019.
- Leslie Pack Kaelbling. Hierarchical Learning in Stochastic Domains: Preliminary Results. In ICML, 1993.
- Language-driven representation learning for robotics. In RSS, 2023.
- From Skills to Symbols: Learning Symbolic Representations for Abstract High-Level Planning. JAIR, 61:215–289, 2018.
- Towards a Unified Theory of State Abstraction for MDPs. In AI&M, 2006.
- Structformer: Learning spatial structure for language-guided semantic rearrangement of novel objects. In International Conference on Robotics and Automation, 2022.
- A survey of reinforcement learning informed by natural language. In IJCAI, 2019.
- Learning rational subgoals from demonstrations and instructions. In Proceedings of the AAAI Conference on Artificial Intelligence, 2023.
- Liv: Language-image representations and rewards for robotic control. In ICML, 2023.
- Pdsketch: Integrated domain programming, learning, and planning. Advances in Neural Information Processing Systems, 35:36972–36984, 2022.
- What matters in language conditioned robotic imitation learning over unstructured data. IEEE Robotics and Automation Letters, 7(4):11205–11212, 2022.
- Neville Mehta. Hierarchical structure discovery and transfer in sequential decision problems. Oregon State University, 2011.
- Listen, attend, and walk: Neural mapping of navigational instructions to action sequences. In Proceedings of the AAAI Conference on Artificial Intelligence, 2016.
- Mapping instructions and visual observations to actions with reinforcement learning. arXiv preprint arXiv:1704.08795, 2017.
- Asynchronous methods for deep reinforcement learning. In International conference on machine learning, 2016.
- Learning language-conditioned robot behavior from offline data and crowd-sourced annotation. In Conference on Robot Learning, pp. 1303–1315. PMLR, 2022.
- Learning macro-actions for arbitrary planners and domains. In ICAPS, 2007.
- Inferring task goals and constraints using bayesian nonparametric inverse reinforcement learning. In JMLR, 2020.
- Learning Symbolic Models of Stochastic Domains. JAIR, 29:309–352, 2007.
- FiLM: Visual Reasoning with a General Conditioning Layer. In AAAI, 2018.
- Learning transferable visual models from natural language supervision. In ICML, 2021.
- Aggregation and Disaggregation Techniques and Methodology in Optimization. Operations Research, 39(4):553–582, 1991.
- Object scene representation transformer. In Advances in Neural Information Processing Systems, 2022.
- Mastering Atari, Go, Chess and Shogi by Planning with a Learned Model. Nat., 588(7839):604–609, 2020.
- Value function spaces: Skill-centric state abstractions for long-horizon reasoning. In ICLR, 2022.
- Skill induction and planning with latent language. arXiv preprint arXiv:2110.01517, 2021.
- Learning Symbolic Operators for Task and Motion Planning. In IROS, 2021.
- Program guided agent. In ICLR, 2020.
- Between MDPs and semi-MDPs: A Framework for Temporal Abstraction in Reinforcement Learning. Artif. Intell., 112(1-2):181–211, 1999.
- Understanding natural language commands for robotic navigation and mobile manipulation. In AAAI, 2011.
- Robots that use language. Annual Review of Control, Robotics, and Autonomous Systems, 3:25–55, 2020.
- Teaching multiple tasks to an rl agent using ltl. In AAMAS, 2018.
- Generalizable task planning through representation pretraining. IEEE Robotics and Automation Letters, 7(3):8299–8306, 2022.
- Programmatically Grounded, Compositionally Generalizable Robotic Manipulation. In ICLR, 2023.
- Language-mediated, object-centric representation learning. In ACL Findings, 2021.
- Regression planning networks. Advances in Neural Information Processing Systems, 32, 2019.
- Sequence-Based Plan Feasibility Prediction for Efficient Task and Motion Planning. In RSS, 2023.
- Piglet: Language grounding through neuro-symbolic interaction in a 3d world. In ACL, 2021.
- Learning Invariant Representations for Reinforcement Learning without Reconstruction. In ICLR, 2021.
- Glipv2: Unifying localization and vision-language understanding. In Advances in Neural Information Processing Systems, 2022.
- Weiyu Liu (22 papers)
- Geng Chen (115 papers)
- Joy Hsu (15 papers)
- Jiayuan Mao (55 papers)
- Jiajun Wu (249 papers)