Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

On the Prospects of Incorporating Large Language Models (LLMs) in Automated Planning and Scheduling (APS) (2401.02500v2)

Published 4 Jan 2024 in cs.AI
On the Prospects of Incorporating Large Language Models (LLMs) in Automated Planning and Scheduling (APS)

Abstract: Automated Planning and Scheduling is among the growing areas in AI where mention of LLMs has gained popularity. Based on a comprehensive review of 126 papers, this paper investigates eight categories based on the unique applications of LLMs in addressing various aspects of planning problems: language translation, plan generation, model construction, multi-agent planning, interactive planning, heuristics optimization, tool integration, and brain-inspired planning. For each category, we articulate the issues considered and existing gaps. A critical insight resulting from our review is that the true potential of LLMs unfolds when they are integrated with traditional symbolic planners, pointing towards a promising neuro-symbolic approach. This approach effectively combines the generative aspects of LLMs with the precision of classical planning methods. By synthesizing insights from existing literature, we underline the potential of this integration to address complex planning challenges. Our goal is to encourage the ICAPS community to recognize the complementary strengths of LLMs and symbolic planners, advocating for a direction in automated planning that leverages these synergistic capabilities to develop more advanced and intelligent planning systems.

Introduction to LLMs and APS

Automated Planning and Scheduling (APS) is a valuable domain in AI, tasked with generating strategies or action sequences for achieving specific goals. Rooted in algorithms and system development, APS automates complex tasks, making systems more intelligent and adaptable. The rise of LLMs in AI, particularly within computational linguistics, has created an unprecedented opportunity to innovate in APS. The focus of the analyzed paper is on the intersection of these two areas, offering a new vantage point in how intelligently systems can plan and schedule tasks by harnessing natural language capabilities.

The Growth of LLMs in APS

LLMs have made significant strides, evolving from basic natural language processing tasks to complex, context-aware text generation. As they become more proficient, these models are increasingly incorporated into APS, using language constructs to define planning elements like preconditions and effects. By integrating traditional symbolic planners with the generative capacity of LLMs, systems can address complex planning challenges with both the creativity of human-like language processing and the accuracy of established planning methods.

Insights from the Literature Review

This paper takes an exhaustive look at recent literature—126 papers on LLMs' role in APS, categorized into eight applications of LLMs in APS: Language Translation, Plan Generation, Model Construction, Multi-agent Planning, Interactive Planning, Heuristics Optimization, Tool Integration, and Brain-Inspired Planning. Each category has been reviewed for the issues addressed and the gaps present. The literature suggests that while LLMs hold potential, their current application is absent of generating action sequences to rival symbolic planners. They shine in scenarios that aren't inherently complex, allowing them to speed up the plan generation process more efficiently than their symbolic counterparts.

Future Directions and Conclusion

The direction for future research is clear. Researchers are encouraged to pursue the development of LLM training methods to improve coherence and goal-oriented outputs and to explore neuro-symbolic integration following suggested taxonomies. Moreover, it's vital to create performance metrics for planners augmented by LLMs. In closing, while LLMs present challenges in their current form, they offer a promising frontier for planning and scheduling. Melding the creative and heuristic advantages of LLM with the exactitude of symbolic approaches stands to propel AI capabilities further into a field that simulates more closely to human reasoning.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (138)
  1. Llm-deliberation: Evaluating llms with interactive multi-agent negotiation games. arXiv preprint arXiv:2309.17234.
  2. Learning and Leveraging Verifiers to Improve Planning Capabilities of Pre-trained Language Models. arXiv preprint arXiv:2305.17077.
  3. Asai, M. 2018. Photo-Realistic Blocksworld Dataset. arXiv preprint arXiv:1812.01818.
  4. Graph of thoughts: Solving elaborate problems with large language models. arXiv preprint arXiv:2308.09687.
  5. Do as i can, not as i say: Grounding language in robotic affordances. In Conference on Robot Learning, 287–318. PMLR.
  6. A Framework to Generate Neurosymbolic PDDL-compliant Planners. arXiv preprint arXiv:2303.00438.
  7. Grounding large language models in interactive environments with online reinforcement learning. arXiv preprint arXiv:2302.02662.
  8. Learning to reason over scene graphs: a case study of finetuning GPT-2 into a robot language model for grounded task planning. Frontiers in Robotics and AI, 10.
  9. Open-vocabulary queryable scene representations for real world planning. In 2023 IEEE International Conference on Robotics and Automation (ICRA), 11509–11522. IEEE.
  10. Asking Before Action: Gather Information in Embodied Decision Making with Language Models. arXiv preprint arXiv:2305.15695.
  11. AutoTAMP: Autoregressive Task and Motion Planning with LLMs as Translators and Checkers. arXiv preprint arXiv:2306.06531.
  12. Scalable Multi-Robot Collaboration with Large Language Models: Centralized or Decentralized Systems? arXiv preprint arXiv:2309.15943.
  13. Dynamic Planning with a LLM. arXiv preprint arXiv:2308.06391.
  14. Optimal Scene Graph Planning with Large Language Model Guidance. arXiv preprint arXiv:2309.09182.
  15. Plan-Seq-Learn: Language Model Guided RL for Solving Long Horizon Robotics Tasks. In 2nd Workshop on Language and Robot Learning: Language as Grounding.
  16. Emergent Cooperation and Strategy Adaptation in Multi-Agent Systems: An Extended Coevolutionary Theory with LLMs. Electronics, 12(12): 2722.
  17. Integrating Action Knowledge and LLMs for Task Planning and Situation Handling in Open Worlds. arXiv preprint arXiv:2305.17590.
  18. Leveraging Commonsense Knowledge from Large Language Models for Task and Motion Planning. In RSS 2023 Workshop on Learning for Task and Motion Planning.
  19. Task and motion planning with large language models for object rearrangement. arXiv preprint arXiv:2303.06247.
  20. Palm-e: An embodied multimodal language model. arXiv preprint arXiv:2303.03378.
  21. From Static to Dynamic: A Continual Learning Framework for Large Language Models. arXiv preprint arXiv:2310.14248.
  22. Halo: Estimation and reduction of hallucinations in open-source weak large language models. arXiv preprint arXiv:2308.11764.
  23. Fast and Slow Planning. arXiv preprint arXiv:2303.04283.
  24. Alphazero-like tree-search can guide large language model decoding and training. arXiv preprint arXiv:2309.17179.
  25. Strategic Reasoning with Language Models. arXiv preprint arXiv:2305.19165.
  26. Openagi: When llm meets domain experts. arXiv preprint arXiv:2304.04370.
  27. HTN planning: Overview, comparison, and beyond. Artif. Intell., 222: 124–156.
  28. LPG: A Planner Based on Local Search for Planning Graphs with Action Costs. In Aips, volume 2, 281–290.
  29. Automated Planning: Theory and Practice. The Morgan Kaufmann Series in Artificial Intelligence. Amsterdam: Morgan Kaufmann. ISBN 978-1-55860-856-6.
  30. Exploring the Limitations of using Large Language Models to Fix Planning Tasks.
  31. Generating executable action plans with environmentally-aware language models. arXiv preprint arXiv:2210.04964.
  32. GG-LLM: Geometrically Grounding Large Language Models for Zero-shot Human Activity Forecasting in Human-Aware Task Planning. arXiv preprint arXiv:2310.20034.
  33. Conceptgraphs: Open-vocabulary 3d scene graphs for perception and planning. arXiv preprint arXiv:2309.16650.
  34. Leveraging Pre-trained Large Language Models to Construct and Utilize World Models for Model-based Task Planning. arXiv preprint arXiv:2305.14909.
  35. Reasoning with language model is planning with world model. arXiv preprint arXiv:2305.14992.
  36. ToolkenGPT: Augmenting Frozen Language Models with Massive Tools via Tool Embeddings. arXiv preprint arXiv:2305.11554.
  37. SayCanPay: Heuristic Planning with Large Language Models using Learnable Domain Knowledge. arXiv preprint arXiv:2308.12682.
  38. VAL: automatic plan validation, continuous effects and mixed initiative planning using PDDL. In 16th IEEE International Conference on Tools with Artificial Intelligence, 294–301.
  39. Tool documentation enables zero-shot tool-usage with large language models. arXiv preprint arXiv:2308.00675.
  40. Enabling Efficient Interaction between an Algorithm Agent and an LLM: A Reinforcement Learning Approach. arXiv preprint arXiv:2306.03604.
  41. Chain-of-Symbol Prompting Elicits Planning in Large Langauge Models. arXiv preprint arXiv:2305.10276.
  42. Tree-Planner: Efficient Close-loop Task Planning with Large Language Models. arXiv preprint arXiv:2310.08582.
  43. Tree-of-mixed-thought: Combining fast and slow thinking for multi-hop visual reasoning. arXiv preprint arXiv:2308.09658.
  44. Language Models, Agent Models, and World Models: The LAW for Machine Reasoning and Planning. arXiv:2312.05230.
  45. Language models as zero-shot planners: Extracting actionable knowledge for embodied agents. In International Conference on Machine Learning, 9118–9147. PMLR.
  46. Voxposer: Composable 3d value maps for robotic manipulation with language models. arXiv preprint arXiv:2307.05973.
  47. Grounded decoding: Guiding text generation with grounded models for robot control. arXiv preprint arXiv:2303.00855.
  48. Inner monologue: Embodied reasoning through planning with language models. arXiv preprint arXiv:2207.05608.
  49. Neuro Symbolic Reasoning for Planning: Counterexample Guided Inductive Synthesis using Large Language Models and Satisfiability Solving. arXiv preprint arXiv:2309.16436.
  50. CoPAL: Corrective Planning of Robot Actions with Large Language Models. arXiv preprint arXiv:2310.07263.
  51. SMART-LLM: Smart Multi-Agent Robot Task Planning using Large Language Models. arXiv preprint arXiv:2309.10062.
  52. Housekeep: Tidying virtual households using commonsense reasoning. In European Conference on Computer Vision, 355–373. Springer.
  53. Kautz, H. A. 2022. The third AI summer: AAAI Robert S. Engelmore Memorial Lecture. AI Magazine, 43(1): 105–125.
  54. There and back again: extracting formal domains for controllable neurosymbolic story authoring. In Proceedings of the AAAI Conference on Artificial Intelligence and Interactive Digital Entertainment, volume 19, 64–74.
  55. DynaCon: Dynamic Robot Planner with Contextual Awareness via LLMs. arXiv preprint arXiv:2309.16031.
  56. Exploiting Language Models as a Source of Knowledge for Cognitive Agents. arXiv preprint arXiv:2310.06846.
  57. Getting from generative ai to trustworthy ai: What llms might learn from cyc. arXiv preprint arXiv:2308.04445.
  58. Api-bank: A benchmark for tool-augmented llms. arXiv preprint arXiv:2304.08244.
  59. Human-Centered Planning. arXiv preprint arXiv:2311.04403.
  60. SwiftSage: A Generative Agent with Fast and Slow Thinking for Complex Interactive Tasks. arXiv preprint arXiv:2305.17390.
  61. Videodirectorgpt: Consistent multi-scene video generation via llm-guided planning. arXiv preprint arXiv:2309.15091.
  62. Text2motion: From natural language instructions to feasible plans. arXiv preprint arXiv:2303.12153.
  63. Llm+ p: Empowering large language models with optimal planning proficiency. arXiv preprint arXiv:2304.11477.
  64. Reflect: Summarizing robot experiences for failure explanation and correction. arXiv preprint arXiv:2306.15724.
  65. Code Models are Zero-shot Precondition Reasoners. arXiv preprint arXiv:2311.09601.
  66. Chameleon: Plug-and-play compositional reasoning with large language models. arXiv preprint arXiv:2304.09842.
  67. Neuro-symbolic causal language planning with commonsense prompting. arXiv e-prints, arXiv–2206.
  68. Multimodal Procedural Planning via Dual Text-Image Prompting. arXiv preprint arXiv:2305.01795.
  69. Reasoning on graphs: Faithful and interpretable large language model reasoning. arXiv preprint arXiv:2310.01061.
  70. Roco: Dialectic multi-robot collaboration with large language models. arXiv preprint arXiv:2307.04738.
  71. PDDL-the planning domain definition language.
  72. Evaluating Cognitive Maps and Planning in Large Language Models with CogEval. arXiv preprint arXiv:2309.15129.
  73. Diversity of Thought Improves Reasoning Abilities of Large Language Models. arXiv preprint arXiv:2310.07088.
  74. Do embodied agents dream of pixelated sheep?: Embodied decision making using language guided world modelling. arXiv preprint arXiv:2301.12050.
  75. Plansformer: Generating symbolic plans using transformers. arXiv preprint arXiv:2212.08681.
  76. Understanding the Capabilities of Large Language Models for Automated Planning. arXiv preprint arXiv:2305.16151.
  77. Plansformer Tool: Demonstrating Generation of Symbolic Plans Using Transformers. In IJCAI, volume 2023, 7158–7162. International Joint Conferences on Artificial Intelligence.
  78. Logic-lm: Empowering large language models with symbolic solvers for faithful logical reasoning. arXiv preprint arXiv:2305.12295.
  79. Human-Assisted Continual Robot Learning with Foundation Models. arXiv preprint arXiv:2309.14321.
  80. Improving Generalization in Task-oriented Dialogues with Workflows and Action Plans. arXiv preprint arXiv:2306.01729.
  81. Saynav: Grounding large language models for dynamic planning to navigation in new environments. arXiv preprint arXiv:2309.04077.
  82. Planning with large language models via corrective re-prompting. In NeurIPS 2022 Foundation Models for Decision Making Workshop.
  83. Sayplan: Grounding large language models using 3d scene graphs for scalable task planning. arXiv preprint arXiv:2307.06135.
  84. Robots that ask for help: Uncertainty alignment for large language model planners. arXiv preprint arXiv:2307.01928.
  85. Tptu: Task planning and tool usage of large language model-based ai agents. arXiv preprint arXiv:2308.03427.
  86. Artificial Intelligence, A Modern Approach. Second Edition.
  87. From Cooking Recipes to Robot Task Trees–Improving Planning Correctness and Task Efficiency by Leveraging LLMs with a Knowledge Network. arXiv preprint arXiv:2309.09181.
  88. Evaluation of Pretrained Large Language Models in Embodied Planning Tasks. In International Conference on Artificial General Intelligence, 222–232. Springer.
  89. RoboVQA: Multimodal Long-Horizon Reasoning for Robotics. arXiv preprint arXiv:2311.00899.
  90. Navigation with large language models: Semantic guesswork as a heuristic for planning. arXiv preprint arXiv:2310.10103.
  91. Hugginggpt: Solving ai tasks with chatgpt and its friends in huggingface. arXiv preprint arXiv:2303.17580.
  92. Vision-Language Interpreter for Robot Task Planning. arXiv preprint arXiv:2311.00967.
  93. Generalized Planning in PDDL Domains with Pretrained Large Language Models. arXiv preprint arXiv:2305.11014.
  94. PDDL planning with pretrained large language models. In NeurIPS 2022 Foundation Models for Decision Making Workshop.
  95. ProgPrompt: program generation for situated robot task planning using large language models. Autonomous Robots, 1–14.
  96. Llm-planner: Few-shot grounded planning for embodied agents with large language models. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2998–3009.
  97. Cognitive architectures for language agents. arXiv preprint arXiv:2309.02427.
  98. AdaPlanner: Adaptive Planning from Feedback with Language Models. arXiv preprint arXiv:2305.16653.
  99. Large Language Models are In-Context Semantic Reasoners rather than Symbolic Reasoners. arXiv preprint arXiv:2305.14825.
  100. Can Large Language Models Really Improve by Self-critiquing Their Own Plans? arXiv preprint arXiv:2310.08118.
  101. PlanBench: An Extensible Benchmark for Evaluating Large Language Models on Planning and Reasoning about Change. In Thirty-seventh Conference on Neural Information Processing Systems Datasets and Benchmarks Track.
  102. Large Language Models Still Can’t Plan (A Benchmark for LLMs on Planning and Reasoning about Change). arXiv preprint arXiv:2206.10498.
  103. On the planning abilities of large language models (a critical investigation with a proposed benchmark). arXiv preprint arXiv:2302.06706.
  104. Conformal Temporal Logic Planning using Large Language Models: Knowing When to Do What and When to Ask for Help. arXiv preprint arXiv:2309.10092.
  105. Plan-and-solve prompting: Improving zero-shot chain-of-thought reasoning by large language models. arXiv preprint arXiv:2305.04091.
  106. Guiding language model reasoning with planning tokens. arXiv preprint arXiv:2310.05707.
  107. Describe, explain, plan and select: Interactive planning with large language models enables open-world multi-task agents. arXiv preprint arXiv:2302.01560.
  108. A Prefrontal Cortex-inspired Architecture for Planning in Large Language Models. arXiv preprint arXiv:2310.00194.
  109. Unleashing the Power of Graph Learning through LLM-based Autonomous Agents. arXiv preprint arXiv:2309.04565.
  110. From Word Models to World Models: Translating from Natural Language to the Probabilistic Language of Thought. arXiv preprint arXiv:2306.12672.
  111. Plan, Eliminate, and Track–Language Models are Good Teachers for Embodied Agents. arXiv preprint arXiv:2305.02412.
  112. Integrating Common Sense and Planning with Large Language Models for Room Tidying. In RSS 2023 Workshop on Learning for Task and Motion Planning.
  113. Embodied task planning with large language models. arXiv preprint arXiv:2307.01848.
  114. Translating natural language to planning goals with large-language models. arXiv preprint arXiv:2302.05128.
  115. Gentopia: A collaborative platform for tool-augmented llms. arXiv preprint arXiv:2308.04030.
  116. Xu, L. 1995. Case based reasoning. IEEE Potentials, 13(5): 10–13.
  117. Creative Robot Tool Use with Large Language Models. arXiv preprint arXiv:2310.13065.
  118. LLM-Grounder: Open-Vocabulary 3D Visual Grounding with Large Language Model as an Agent. arXiv preprint arXiv:2309.12311.
  119. OceanChat: Piloting Autonomous Underwater Vehicles in Natural Language. arXiv preprint arXiv:2309.16052.
  120. Learning Automata-Based Task Knowledge Representation from Large-Scale Generative Language Models. arXiv preprint arXiv:2212.01944.
  121. On the Planning, Search, and Memorization Capabilities of Large Language Models. arXiv preprint arXiv:2309.01868.
  122. Yang, Z. 2023. Neuro-Symbolic AI Approaches to Enhance Deep Neural Networks with Logical Reasoning and Knowledge Integration. Ph.D. thesis, Arizona State University.
  123. Coupling Large Language Models with Logic Programming for Robust and General Reasoning from Text. arXiv preprint arXiv:2307.07696.
  124. Tree of thoughts: Deliberate problem solving with large language models. arXiv preprint arXiv:2305.10601.
  125. Statler: State-maintaining language models for embodied reasoning. arXiv preprint arXiv:2306.17840.
  126. EIPE-text: Evaluation-Guided Iterative Plan Extraction for Long-Form Narrative Text Generation. arXiv preprint arXiv:2310.08185.
  127. Plan4mc: Skill reinforcement learning and planning for open-world minecraft tasks. arXiv preprint arXiv:2303.16563.
  128. Distilling Script Knowledge from Large Language Models for Constrained Language Planning. arXiv preprint arXiv:2305.05252.
  129. Parsel: Algorithmic Reasoning with Language Models by Composing Decompositions. In Thirty-seventh Conference on Neural Information Processing Systems.
  130. Large language models as zero-shot human models for human-robot interaction. arXiv preprint arXiv:2303.03548.
  131. Prefer: Prompt ensemble learning via feedback-reflect-refine. arXiv preprint arXiv:2308.12033.
  132. Planning with Logical Graph-based Language Model for Instruction Generation. arXiv:2308.13782.
  133. Building cooperative embodied agents modularly with large language models. arXiv preprint arXiv:2307.02485.
  134. Bootstrap your own skills: Learning to solve new tasks with large language model guidance. arXiv preprint arXiv:2310.10021.
  135. Large Language Models as Commonsense Knowledge for Large-Scale Task Planning. arXiv preprint arXiv:2305.14078.
  136. Judging LLM-as-a-judge with MT-Bench and Chatbot Arena. arXiv:2306.05685.
  137. Steve-Eye: Equipping LLM-based Embodied Agents with Visual Perception in Open Worlds. arXiv preprint arXiv:2310.13255.
  138. ISR-LLM: Iterative Self-Refined Large Language Model for Long-Horizon Sequential Task Planning. arXiv preprint arXiv:2308.13724.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Vishal Pallagani (17 papers)
  2. Kaushik Roy (265 papers)
  3. Bharath Muppasani (9 papers)
  4. Francesco Fabiano (16 papers)
  5. Andrea Loreggia (20 papers)
  6. Keerthiram Murugesan (38 papers)
  7. Biplav Srivastava (57 papers)
  8. Francesca Rossi (55 papers)
  9. Lior Horesh (52 papers)
  10. Amit Sheth (127 papers)
Citations (22)