VAL: Interactive Task Learning with GPT Dialog Parsing (2310.01627v2)
Abstract: Machine learning often requires millions of examples to produce static, black-box models. In contrast, interactive task learning (ITL) emphasizes incremental knowledge acquisition from limited instruction provided by humans in modalities such as natural language. However, ITL systems often suffer from brittle, error-prone language parsing, which limits their usability. LLMs are resistant to brittleness but are not interpretable and cannot learn incrementally. We present VAL, an ITL system with a new philosophy for LLM/symbolic integration. By using LLMs only for specific tasks--such as predicate and argument selection--within an algorithmic framework, VAL reaps the benefits of LLMs to support interactive learning of hierarchical task knowledge from natural language. Acquired knowledge is human interpretable and generalizes to support execution of novel tasks without additional training. We studied users' interactions with VAL in a video game setting, finding that most users could successfully teach VAL using language they felt was natural.
- SHOP2: An HTN Planning System. J. Artif. Intell. Res. 20 (2003), 379–404. https://api.semanticscholar.org/CorpusID:1914873
- On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency (Virtual Event, Canada) (FAccT ’21). Association for Computing Machinery, New York, NY, USA, 610–623. https://doi.org/10.1145/3442188.3445922
- Language models are few-shot learners. Advances in neural information processing systems 33 (2020), 1877–1901.
- On the Utility of Learning about Humans for Human-AI Coordination. In Proceedings of the 33rd International Conference on Neural Information Processing Systems. Curran Associates, 1055 Canada Pl, Vancouver, BC V6C 0C3, Canada, 5175–5186.
- On the Opportunities and Risks of Foundation Models. arXiv:2108.07258 [cs.LG]
- BLOOM: A 176B-Parameter Open-Access Multilingual Language Model. arXiv:2211.05100 [cs.CL]
- Ghost Town Games. 2016. Overcooked. https://store.steampowered.com/app/448510/Overcooked/
- Scaling Laws for Reward Model Overoptimization. arXiv:2210.10760 [cs.LG]
- Sajad Hashemian. 2023. Simple Chat UI. https://codepen.io/sajadhsm/pen/odaBdd.
- David R. Hill. 1984. Dialogue design notes. Research Report 84/167/25. The University of Calgary. https://pages.cpsc.ucalgary.ca/~hill/papers/dialogue-design-notes.pdf
- Scott B. Huffman and John E. Laird. 1993. Learning Procedures from Interactive Natural Language Instructions. In Machine Learning, Proceedings of the Tenth International Conference, University of Massachusetts, Amherst, MA, USA, June 27-29, 1993, Paul E. Utgoff (Ed.). Morgan Kaufmann, Amherst, MA 01003, 143–150. https://doi.org/10.1016/b978-1-55860-307-3.50025-3
- James R. Kirk and John E. Laird. 2014. Interactive Task Learning for Simple Games. Advances in Cognitive Systems 3, 13-30 (2014), 5.
- James R. Kirk and John E. Laird. 2019. Learning Hierarchical Symbolic Representations to Support Interactive Task Learning and Knowledge Transfer. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI-19. International Joint Conferences on Artificial Intelligence Organization, Macao, China, 6095–6102. https://doi.org/10.24963/ijcai.2019/844
- Interactive Task Learning. IEEE Intelligent Systems 32, 4 (2017), 6–21.
- Soar: An Architecture for General Intelligence. Artificial intelligence 33, 1 (1987), 1–64.
- Learning Probabilistic Hierarchical Task Networks as Probabilistic Context-Free Grammars to Capture User Preferences. ACM Trans. Intell. Syst. Technol. 5, 2, Article 29 (apr 2014), 32 pages. https://doi.org/10.1145/2589481
- PUMICE: A Multi-Modal Agent That Learns Concepts and Conditionals from Natural Language and Demonstrations. In Proceedings of the 32nd Annual ACM Symposium on User Interface Software and Technology (New Orleans, LA, USA) (UIST ’19). Association for Computing Machinery, New York, NY, USA, 577–589. https://doi.org/10.1145/3332165.3347899
- A framework for natural cognitive system training interactions. Advances in Cognitive Systems 6 (2018), 1–16.
- A hierarchical task network planner for pathfinding in real-time strategy games. In Proceedings of the 3rd International Symposium on AI and Games - A Symposium at the AISB 2010 Convention. Society for the Study of Artificial Intelligence and Simulation of Behaviour, Gateway House, Leicester LE1 9BH, United Kingdom, 1–7. 3rd International Symposium on Artificial Intelligence and Games - A Symposium at the AISB 2010 Convention ; Conference date: 29-03-2010 Through 01-04-2010.
- Santiago Ontañón and Michael Buro. 2015. Adversarial Hierarchical-Task Network Planning for Complex Real-Time Games. In Proceedings of the 24th International Conference on Artificial Intelligence (IJCAI’15). AAAI Press, Buenos Aires, Argentina, 1652–1658.
- OpenAI. 2023. GPT-4 Technical Report. arXiv:2303.08774 [cs.CL]
- Sentence Similarity based on Dependency Tree Kernels for Multi-document Summarization. In Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC’16). European Language Resources Association (ELRA), Portorož, Slovenia, 2833–2838. https://aclanthology.org/L16-1452
- Panupong Pasupat and Percy Liang. 2015. Compositional Semantic Parsing on Semi-Structured Tables. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers). Association for Computational Linguistics, Beijing, China, 1470–1480. https://doi.org/10.3115/v1/P15-1142
- ONYX - User Interfaces for Assisting in Interactive Task Learning for Natural Language Interfaces of Data Visualization Tools. In Extended Abstracts of the 2022 CHI Conference on Human Factors in Computing Systems (New Orleans, LA, USA) (CHI EA ’22). Association for Computing Machinery, New York, NY, USA, Article 433, 7 pages. https://doi.org/10.1145/3491101.3519793
- Earl D. Sacerdoti. 1975. The Nonlinear Nature of Plans. In Proceedings of the 4th International Joint Conference on Artificial Intelligence - Volume 1 (Tblisi, USSR) (IJCAI’75). Morgan Kaufmann Publishers Inc., San Francisco, CA, USA, 206–214.
- R. Sheline and C.J. MacLellan. 2018. Investigating Machine-Learning Interaction with Wizard-of-Oz Experiments. In Proceedings of the NeurIPS 2018 Workshop on Learning by Instruction. Curran Associates, Montreal, QC, Canada, 8 pages.
- Teaching Robots Generalizable Hierarchical Tasks Through Natural Language Instruction. IEEE Robotics and Automation Letters 2, 1 (2017), 201–208. https://doi.org/10.1109/LRA.2016.2588584
- BERT Rediscovers the Classical NLP Pipeline. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, Florence, Italy, 4593–4601. https://doi.org/10.18653/v1/P19-1452
- LLaMA: Open and Efficient Foundation Language Models. arXiv:2302.13971 [cs.CL]
- Towards Mutual Theory of Mind in Human-AI Interaction: How Language Reflects What Students Perceive About a Virtual Teaching Assistant. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems (Yokohama, Japan) (CHI ’21). Association for Computing Machinery, New York, NY, USA, Article 384, 14 pages. https://doi.org/10.1145/3411764.3445645
- Fast and easy language understanding for dialog systems with Microsoft Language Understanding Intelligent Service (LUIS). In Proceedings of the 16th Annual Meeting of the Special Interest Group on Discourse and Dialogue. Association for Computational Linguistics, Prague, Czech Republic, 159–161. https://doi.org/10.18653/v1/W15-4622
- Xiaocheng Yang and Anca Dragan. 2021. Enabling Generalization of Human Models for Human-AI Collaboration to New Tasks. Master’s thesis. EECS Department, University of California, Berkeley. http://www2.eecs.berkeley.edu/Pubs/TechRpts/2021/EECS-2021-199.html
- COACH: Cooperative Robot Teaching. In Conference on Robot Learning. PMLR, Conference on Robot Learning, 188 14th Street, Atlanta, Georgia 30361, 1092–1103.
- ProAgent: Building Proactive Cooperative AI with Large Language Models. arXiv preprint arXiv:2308.11339 ∞\infty∞, ∞\infty∞ (2023), 22 pages.
- OPT: Open Pre-trained Transformer Language Models. arXiv:2205.01068 [cs.CL]