Learning adaptive planning representations with natural language guidance (2312.08566v1)
Abstract: Effective planning in the real world requires not only world knowledge, but the ability to leverage that knowledge to build the right representation of the task at hand. Decades of hierarchical planning techniques have used domain-specific temporal action abstractions to support efficient and accurate planning, almost always relying on human priors and domain knowledge to decompose hard tasks into smaller subproblems appropriate for a goal or set of goals. This paper describes Ada (Action Domain Acquisition), a framework for automatically constructing task-specific planning representations using task-general background knowledge from LLMs (LMs). Starting with a general-purpose hierarchical planner and a low-level goal-conditioned policy, Ada interactively learns a library of planner-compatible high-level action abstractions and low-level controllers adapted to a particular domain of planning tasks. On two language-guided interactive planning benchmarks (Mini Minecraft and ALFRED Household Tasks), Ada strongly outperforms other approaches that use LMs for sequential decision-making, offering more accurate plans and better generalization to complex tasks.
- Do as I Can, Not as I Say: Grounding Language in Robotic Affordances. arXiv:2204.01691, 2022.
- Hierarchical Planning: Relating Task and Goal Decomposition with Task Sharing. In IJCAI, 2016.
- Modular Multitask Reinforcement Learning with Policy Sketches. In ICML, 2017.
- Hindsight Experience Replay. In NeurIPS, 2017.
- Learning and Leveraging Verifiers to Improve Planning Capabilities of Pre-trained Language Models. arXiv:2305.17077, 2023.
- Learning Neural-Symbolic Descriptive Planning Models via Cube-Space Priors: The Voyage Home (To Strips). In IJCAI, 2020.
- Learning First-Order Symbolic Representations for Planning from the Structure of the State Space. In ECAI, 2020.
- Top-Down Synthesis for Library Learning. PACMPL, 7(POPL):1182–1213, 2023.
- Babble: Learning Better Abstractions with E-graphs and Anti-unification. PACMPL, 7(POPL):396–424, 2023.
- Ask Your Humans: Using Human Instructions to Improve Generalization in Reinforcement Learning. In ICLR, 2021.
- GLiB: Efficient Exploration for Relational Model-Based Reinforcement Learning via Goal-Literal Babbling. In AAAI, 2021.
- Modular Networks for Compositional Instruction Following. In NAACL-HLT, 2021.
- Tomás de la Rosa and Sheila McIlraith. Learning Domain Control Knowledge for TLPlan and Beyond. In ICAPS 2011 Workshop on Planning and Learning, 2011.
- Thomas G Dietterich. Hierarchical Reinforcement Learning with the MAXQ Value Function Decomposition. JAIR, 13:227–303, 2000.
- DreamCoder: Growing Generalizable, Interpretable Knowledge with Wake–Sleep Bayesian Program Learning. Philosophical Transactions of the Royal Society, 381(2251):20220050, 2023.
- HTN Planning: Complexity and Expressivity. In AAAI, 1994.
- STRIPS: A New Approach to the Application of Theorem Proving to Problem Solving. Artif. Intell., 2(3-4):189–208, 1971.
- Integrated Task and Motion Planning. Ann. Rev. Control Robot. Auton. Syst., 4:265–293, 2021.
- Malte Helmert. The Fast Downward Planning System. JAIR, 26:191–246, 2006.
- People construct simplified mental representations to plan. Nature, 606(7912):129–136, 2022.
- Language Models as Zero-Shot Planners: Extracting Actionable Knowledge for Embodied Agents. In ICML, 2022.
- Language as an Abstraction for Hierarchical Deep Reinforcement Learning. In NeurIPS, 2019.
- From Skills to Symbols: Learning Symbolic Representations for Abstract High-Level Planning. JAIR, 61:215–289, 2018.
- Steven LaValle. Rapidly-Exploring Random Trees: A New Tool for Path Planning. Research Report 9811, 1998.
- Vladimir Lifschitz. On the Semantics of STRIPS. In Workshop on Reasoning about Actions and Plans, 1986.
- LLM+ P: Empowering Large Language Models with Optimal Planning Proficiency. arXiv:2304.11477, 2023.
- A Survey of Reinforcement Learning Informed by Natural Language. In IJCAI, 2019.
- Learning Rational Subgoals from Demonstrations and Instructions. In AAAI, 2023.
- PDSketch: Integrated Domain Programming, Learning, and Planning. In NeurIPS, 2022.
- Grounding predicates through actions, 2022.
- Mapping Instructions and Visual Observations to Actions with Reinforcement Learning. In EMNLP, 2017.
- EmbodiedGPT: Vision-Language Pre-training via Embodied Chain of Thought. arXiv:2305.15021, 2023.
- Learning Language-Conditioned Robot Behavior from Offline Data and Crowd-Sourced Annotation. In CoRL, 2022.
- Learning Hierarchical Task Networks by Observation. In ICML, 2006.
- Do embodied agents dream of pixelated sheep?: Embodied decision making using language guided world modelling. arXiv preprint arXiv:2301.12050, 2023.
- Skill Induction and Planning with Latent Language. In ACL, 2022.
- ALFRED: A Benchmark for Interpreting Grounded Instructions for Everyday Tasks. In CVPR, 2020.
- Learning Symbolic Operators for Task and Motion Planning. In IROS, 2021.
- PDDL Planning with Pretrained Large Language Models. In NeurIPS Foundation Models for Decision Making Workshop, 2022.
- Generalized Planning in PDDL Domains with Pretrained Large Language Models. arXiv:2305.11014, 2023.
- LLM-Planner: Few-Shot Grounded Planning for Embodied Agents with Large Language Models. In ICCV, 2023.
- Between MDPs and Semi-MDPs: A Framework for Temporal Abstraction in Reinforcement Learning. Artif. Intell., 112(1-2):181–211, 1999.
- Understanding Natural Language Commands for Robotic Navigation and Mobile Manipulation. In AAAI, 2011.
- Robots That Use Language. Annual Review of Control, Robotics, & Autonomous Systems, 3:25–55, 2020.
- Large Language Models Still Can’t Plan (A Benchmark for LLMs on Planning and Reasoning about Change). In NeurIPS Foundation Models for Decision Making Workshop, 2022.
- Voyager: An Open-Ended Embodied Agent with Large Language Models. arXiv:2305.16291, 2023a.
- Describe, explain, plan and select: Interactive planning with large language models enables open-world multi-task agents, 2023b.
- Leveraging Language to Learn Program Abstractions and Search Heuristics. In ICML, 2021.
- Translating Natural Language to Planning Goals with Large-Language Models. arXiv:2302.05128, 2023.
- Ghost in the minecraft: Generally capable agents for open-world environments via large language models with text-based knowledge and memory, 2023.