The Case for Developing a Foundation Model for Planning-like Tasks from Scratch (2404.04540v1)
Abstract: Foundation Models (FMs) have revolutionized many areas of computing, including Automated Planning and Scheduling (APS). For example, a recent study found them useful for planning problems: plan generation, language translation, model construction, multi-agent planning, interactive planning, heuristics optimization, tool integration, and brain-inspired planning. Besides APS, there are many seemingly related tasks involving the generation of a series of actions with varying guarantees of their executability to achieve intended goals, which we collectively call planning-like (PL) tasks like business processes, programs, workflows, and guidelines, where researchers have considered using FMs. However, previous works have primarily focused on pre-trained, off-the-shelf FMs and optionally fine-tuned them. This paper discusses the need for a comprehensive FM for PL tasks from scratch and explores its design considerations. We argue that such an FM will open new and efficient avenues for PL problem-solving, just like LLMs are creating for APS.
- Understanding approaches for web service composition and execution. In Proceedings of the 1st Bangalore annual Compute conference, 1–8.
- A dataset to facilitate automated workflow analysis. In PLoS ONE 14(2): e0211486. https://doi.org/10.1371/journal.pone.0211486.
- Generating Dialogue Agents via Automated Planning. In https://arxiv.org/abs/1902.00771.
- Camunda. 2024. BPMN for research. In https://github.com/camunda/bpmn-for-research.
- Web Services Description Language (WSDL) Version 2.0 Part 1: Core Language. World Wide Web Consortium, Recommendation REC-wsdl20-20070626.
- PaLM: Scaling Language Modeling with Pathways. arXiv:2204.02311.
- LLM.int8(): 8-bit Matrix Multiplication for Transformers at Scale. arXiv:2208.07339.
- Fast and Slow Planning. arXiv preprint arXiv:2303.04283.
- Clinical practice guidelines. Directions for a new program, 90(8).
- A Survey of Quantization Methods for Efficient Neural Network Inference. arXiv:2103.13630.
- ImageBind: One Embedding Space To Bind Them All. In CVPR.
- Knowledge Distillation of Large Language Models. arXiv:2306.08543.
- Textbooks are all you need. arXiv preprint arXiv:2306.11644.
- Hannibal046. 2024. Awesome-LLM: a curated list of Large Language Model. In https://github.com/Hannibal046/Awesome-LLM.
- Ethical Challenges in Data-Driven Dialogue Systems. In Proc. of AAAI/ACM Conference on AI Ethics and Society (AIES-18), New Orleans, Lousiana, USA.
- Distilling Step-by-Step! Outperforming Larger Language Models with Less Training Data and Smaller Model Sizes. arXiv:2305.02301.
- HuggingFace. 2024. Training a causal language model from scratch. https://huggingface.co/learn/nlp-course/en/chapter7/6.
- ICAPS. 2023. International Planning Competitions. In https://www.icaps-conference.org/competitions/.
- Chatgpt and software testing education: Promises & perils. In 2023 IEEE International Conference on Software Testing, Verification and Validation Workshops (ICSTW), 4130–4137. IEEE.
- Karpathy, A. 2023. Let’s build GPT: from scratch, in code, spelled out. https://www.youtube.com/watch?v=kCc8FmEb1nY.
- Reference Manual. In The C Programming Language, chapter Appendix A, 191–240. Prentice Hall, second edition.
- Beyond A*: Better Planning with Transformers via Search Dynamics Bootstrapping. arXiv preprint arXiv:2402.14083.
- Retrieval-augmented generation for knowledge-intensive NLP tasks. In Proceedings of the 34th International Conference on Neural Information Processing Systems, NIPS ’20. Red Hook, NY, USA: Curran Associates Inc. ISBN 9781713829546.
- LLM-QAT: Data-Free Quantization Aware Training for Large Language Models. arXiv:2305.17888.
- LLM-Pruner: On the Structural Pruning of Large Language Models. arXiv:2305.11627.
- PDDL-the planning domain definition language.
- Conversational Interfaces: Past and Present. In The Conversational Interface. Springer, DOI: https://doi.org/10.1007/978-3-319-32967-3_4.
- Planning for Goal-Oriented Dialogue Systems. In https://arxiv.org/abs/1910.08137.
- Building and using a planning ontology from past data for performance efficiency. In Proc. ICAPS’23 Workshop PLanning And onTology wOrkshop (PLATO).
- OMG. 2011. Business Process Model and Notation (BPMN), Version 2.0.
- OpenAI. 2023. GPT-4 Technical Report. arXiv:2303.08774.
- Plansformer: Generating symbolic plans using transformers. arXiv preprint arXiv:2212.08681.
- Plansformer: Generating Symbolic Plans using Transformers. In On Arxiv at: https://arxiv.org/abs/2212.08681.
- Understanding the Capabilities of Large Language Models for Automated Planning. arXiv preprint arXiv:2305.16151.
- Plansformer Tool: Demonstrating Generation of Symbolic Plans Using Transformers. In IJCAI, volume 2023, 7158–7162. International Joint Conferences on Artificial Intelligence.
- Harnessing Large Language Models for Planning: A Lab on Strategies for Success and Mitigation of Pitfalls. In AAAI Conference on Artificial Intelligence, https://github.com/VishalPallagani/LLMsforPlanningLab-AAAI24.
- On the Prospects of Incorporating Large Language Models (LLMs) in Automated Planning and Scheduling (APS). In ICAPS, Calgary, Canada.
- A Generic Dialog Agent for Information Retrieval Based on Automated Planning Within a Reinforcement Learning Platform.
- Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. Journal of Machine Learning Research, 21(140): 1–67.
- Exploring the limits of transfer learning with a unified text-to-text transformer. Journal of machine learning research, 21(140): 1–67.
- Artificial Intelligence, A Modern Approach. Second Edition.
- DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. arXiv:1910.01108.
- Process knowledge-infused ai: Toward user-level explainability, interpretability, and safety. IEEE Internet Computing, 26(5): 76–84.
- Srivastava, B. 2010. Processes Summarization. In Conference on Management of Data (COMAD), India. Code at: https://github.com/biplav-s/processes-summarizer.
- Srivastava, B. 2021. Did chatbots miss their “Apollo Moment”? Potential, gaps, and lessons from using collaboration assistants during COVID-19. In Patterns, Volume 2, Issue 8, 100308.
- Web service composition-current solutions and open problems. In ICAPS 2003 workshop on Planning for Web Services, volume 35, 28–35.
- An APQC-PCF based framework to compare service offerings in business transformation projects. In Proceedings of the 2010 ACM Symposium on Applied Computing, 73–78.
- Business Driven Consolidation of SOA Implementations. In 2010 IEEE International Conference on Services Computing, 49–56. IEEE.
- Roformer: Enhanced transformer with rotary position embedding. Neurocomputing, 568: 127063.
- A Simple and Effective Pruning Approach for Large Language Models. arXiv:2306.11695.
- Transfer learning. In Handbook of research on machine learning applications and trends: algorithms, methods, and techniques, 242–264. IGI global.
- LLaMA: Open and Efficient Foundation Language Models. arXiv:2302.13971.
- Uncbiag. 2024. Awesome Foundation Models. In https://github.com/uncbiag/Awesome-Foundation-Models.
- Can Large Language Models Really Improve by Self-critiquing Their Own Plans? arXiv preprint arXiv:2310.08118.
- PlanBench: An Extensible Benchmark for Evaluating Large Language Models on Planning and Reasoning about Change. In Thirty-seventh Conference on Neural Information Processing Systems Datasets and Benchmarks Track.
- Large Language Models Still Can’t Plan (A Benchmark for LLMs on Planning and Reasoning about Change). arXiv preprint arXiv:2206.10498.
- On the planning abilities of large language models (a critical investigation with a proposed benchmark). arXiv preprint arXiv:2302.06706.
- Workflow Management: Models, Methods, and Systems. In ISBN:978-0262720465, MIT Press.
- Attention is all you need. Advances in neural information processing systems, 30.
- A Survey of Available Corpora for Building Data-Driven Dialogue Systems. ArXiv e-prints.
- From Cloud to Edge: Rethinking Generative AI for Low-Resource Design Challenges. arXiv:2402.12702.
- Codet5: Identifier-aware unified pre-trained encoder-decoder models for code understanding and generation. arXiv preprint arXiv:2109.00859.
- Weitzenhoffer, A. M. 1974. When is an “instruction” an “instruction”? International Journal of Clinical and Experimental Hypnosis, 22(3): 258–269.
- DeepCAD: A Deep Generative Network for Computer-Aided Design Models. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 6772–6782.
- Travelplanner: A benchmark for real-world planning with language agents. arXiv preprint arXiv:2402.01622.
- Measuring GitHub Copilot’s Impact on Productivity. Commun. ACM, 67(3): 54–63.
- Fine-tuning language models from human preferences. arXiv preprint arXiv:1909.08593.