Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
158 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

VAL: Interactive Task Learning with GPT Dialog Parsing (2310.01627v2)

Published 2 Oct 2023 in cs.HC, cs.AI, and cs.CL

Abstract: Machine learning often requires millions of examples to produce static, black-box models. In contrast, interactive task learning (ITL) emphasizes incremental knowledge acquisition from limited instruction provided by humans in modalities such as natural language. However, ITL systems often suffer from brittle, error-prone language parsing, which limits their usability. LLMs are resistant to brittleness but are not interpretable and cannot learn incrementally. We present VAL, an ITL system with a new philosophy for LLM/symbolic integration. By using LLMs only for specific tasks--such as predicate and argument selection--within an algorithmic framework, VAL reaps the benefits of LLMs to support interactive learning of hierarchical task knowledge from natural language. Acquired knowledge is human interpretable and generalizes to support execution of novel tasks without additional training. We studied users' interactions with VAL in a video game setting, finding that most users could successfully teach VAL using language they felt was natural.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (35)
  1. SHOP2: An HTN Planning System. J. Artif. Intell. Res. 20 (2003), 379–404. https://api.semanticscholar.org/CorpusID:1914873
  2. On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency (Virtual Event, Canada) (FAccT ’21). Association for Computing Machinery, New York, NY, USA, 610–623. https://doi.org/10.1145/3442188.3445922
  3. Language models are few-shot learners. Advances in neural information processing systems 33 (2020), 1877–1901.
  4. On the Utility of Learning about Humans for Human-AI Coordination. In Proceedings of the 33rd International Conference on Neural Information Processing Systems. Curran Associates, 1055 Canada Pl, Vancouver, BC V6C 0C3, Canada, 5175–5186.
  5. On the Opportunities and Risks of Foundation Models. arXiv:2108.07258 [cs.LG]
  6. BLOOM: A 176B-Parameter Open-Access Multilingual Language Model. arXiv:2211.05100 [cs.CL]
  7. Ghost Town Games. 2016. Overcooked. https://store.steampowered.com/app/448510/Overcooked/
  8. Scaling Laws for Reward Model Overoptimization. arXiv:2210.10760 [cs.LG]
  9. Sajad Hashemian. 2023. Simple Chat UI. https://codepen.io/sajadhsm/pen/odaBdd.
  10. David R. Hill. 1984. Dialogue design notes. Research Report 84/167/25. The University of Calgary. https://pages.cpsc.ucalgary.ca/~hill/papers/dialogue-design-notes.pdf
  11. Scott B. Huffman and John E. Laird. 1993. Learning Procedures from Interactive Natural Language Instructions. In Machine Learning, Proceedings of the Tenth International Conference, University of Massachusetts, Amherst, MA, USA, June 27-29, 1993, Paul E. Utgoff (Ed.). Morgan Kaufmann, Amherst, MA 01003, 143–150. https://doi.org/10.1016/b978-1-55860-307-3.50025-3
  12. James R. Kirk and John E. Laird. 2014. Interactive Task Learning for Simple Games. Advances in Cognitive Systems 3, 13-30 (2014), 5.
  13. James R. Kirk and John E. Laird. 2019. Learning Hierarchical Symbolic Representations to Support Interactive Task Learning and Knowledge Transfer. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI-19. International Joint Conferences on Artificial Intelligence Organization, Macao, China, 6095–6102. https://doi.org/10.24963/ijcai.2019/844
  14. Interactive Task Learning. IEEE Intelligent Systems 32, 4 (2017), 6–21.
  15. Soar: An Architecture for General Intelligence. Artificial intelligence 33, 1 (1987), 1–64.
  16. Learning Probabilistic Hierarchical Task Networks as Probabilistic Context-Free Grammars to Capture User Preferences. ACM Trans. Intell. Syst. Technol. 5, 2, Article 29 (apr 2014), 32 pages. https://doi.org/10.1145/2589481
  17. PUMICE: A Multi-Modal Agent That Learns Concepts and Conditionals from Natural Language and Demonstrations. In Proceedings of the 32nd Annual ACM Symposium on User Interface Software and Technology (New Orleans, LA, USA) (UIST ’19). Association for Computing Machinery, New York, NY, USA, 577–589. https://doi.org/10.1145/3332165.3347899
  18. A framework for natural cognitive system training interactions. Advances in Cognitive Systems 6 (2018), 1–16.
  19. A hierarchical task network planner for pathfinding in real-time strategy games. In Proceedings of the 3rd International Symposium on AI and Games - A Symposium at the AISB 2010 Convention. Society for the Study of Artificial Intelligence and Simulation of Behaviour, Gateway House, Leicester LE1 9BH, United Kingdom, 1–7. 3rd International Symposium on Artificial Intelligence and Games - A Symposium at the AISB 2010 Convention ; Conference date: 29-03-2010 Through 01-04-2010.
  20. Santiago Ontañón and Michael Buro. 2015. Adversarial Hierarchical-Task Network Planning for Complex Real-Time Games. In Proceedings of the 24th International Conference on Artificial Intelligence (IJCAI’15). AAAI Press, Buenos Aires, Argentina, 1652–1658.
  21. OpenAI. 2023. GPT-4 Technical Report. arXiv:2303.08774 [cs.CL]
  22. Sentence Similarity based on Dependency Tree Kernels for Multi-document Summarization. In Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC’16). European Language Resources Association (ELRA), Portorož, Slovenia, 2833–2838. https://aclanthology.org/L16-1452
  23. Panupong Pasupat and Percy Liang. 2015. Compositional Semantic Parsing on Semi-Structured Tables. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers). Association for Computational Linguistics, Beijing, China, 1470–1480. https://doi.org/10.3115/v1/P15-1142
  24. ONYX - User Interfaces for Assisting in Interactive Task Learning for Natural Language Interfaces of Data Visualization Tools. In Extended Abstracts of the 2022 CHI Conference on Human Factors in Computing Systems (New Orleans, LA, USA) (CHI EA ’22). Association for Computing Machinery, New York, NY, USA, Article 433, 7 pages. https://doi.org/10.1145/3491101.3519793
  25. Earl D. Sacerdoti. 1975. The Nonlinear Nature of Plans. In Proceedings of the 4th International Joint Conference on Artificial Intelligence - Volume 1 (Tblisi, USSR) (IJCAI’75). Morgan Kaufmann Publishers Inc., San Francisco, CA, USA, 206–214.
  26. R. Sheline and C.J. MacLellan. 2018. Investigating Machine-Learning Interaction with Wizard-of-Oz Experiments. In Proceedings of the NeurIPS 2018 Workshop on Learning by Instruction. Curran Associates, Montreal, QC, Canada, 8 pages.
  27. Teaching Robots Generalizable Hierarchical Tasks Through Natural Language Instruction. IEEE Robotics and Automation Letters 2, 1 (2017), 201–208. https://doi.org/10.1109/LRA.2016.2588584
  28. BERT Rediscovers the Classical NLP Pipeline. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, Florence, Italy, 4593–4601. https://doi.org/10.18653/v1/P19-1452
  29. LLaMA: Open and Efficient Foundation Language Models. arXiv:2302.13971 [cs.CL]
  30. Towards Mutual Theory of Mind in Human-AI Interaction: How Language Reflects What Students Perceive About a Virtual Teaching Assistant. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems (Yokohama, Japan) (CHI ’21). Association for Computing Machinery, New York, NY, USA, Article 384, 14 pages. https://doi.org/10.1145/3411764.3445645
  31. Fast and easy language understanding for dialog systems with Microsoft Language Understanding Intelligent Service (LUIS). In Proceedings of the 16th Annual Meeting of the Special Interest Group on Discourse and Dialogue. Association for Computational Linguistics, Prague, Czech Republic, 159–161. https://doi.org/10.18653/v1/W15-4622
  32. Xiaocheng Yang and Anca Dragan. 2021. Enabling Generalization of Human Models for Human-AI Collaboration to New Tasks. Master’s thesis. EECS Department, University of California, Berkeley. http://www2.eecs.berkeley.edu/Pubs/TechRpts/2021/EECS-2021-199.html
  33. COACH: Cooperative Robot Teaching. In Conference on Robot Learning. PMLR, Conference on Robot Learning, 188 14th Street, Atlanta, Georgia 30361, 1092–1103.
  34. ProAgent: Building Proactive Cooperative AI with Large Language Models. arXiv preprint arXiv:2308.11339 ∞\infty∞, ∞\infty∞ (2023), 22 pages.
  35. OPT: Open Pre-trained Transformer Language Models. arXiv:2205.01068 [cs.CL]
Citations (6)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets