Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Improving Knowledge Extraction from LLMs for Task Learning through Agent Analysis (2306.06770v4)

Published 11 Jun 2023 in cs.AI, cs.HC, and cs.RO

Abstract: LLMs offer significant promise as a knowledge source for task learning. Prompt engineering has been shown to be effective for eliciting knowledge from an LLM, but alone it is insufficient for acquiring relevant, situationally grounded knowledge for an embodied agent learning novel tasks. We describe a cognitive-agent approach, STARS, that extends and complements prompt engineering, mitigating its limitations and thus enabling an agent to acquire new task knowledge matched to its native language capabilities, embodiment, environment, and user preferences. The STARS approach is to increase the response space of LLMs and deploy general strategies, embedded within the autonomous agent, to evaluate, repair, and select among candidate responses produced by the LLM. We describe the approach and experiments that show how an agent, by retrieving and evaluating a breadth of responses from the LLM, can achieve 77-94% task completion in one-shot learning without user oversight. The approach achieves 100% task completion when human oversight (such as an indication of preference) is provided. Further, the type of oversight largely shifts from explicit, natural language instruction to simple confirmation/discomfirmation of high-quality responses that have been vetted by the agent before presentation to a user.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (27)
  1. Do As I Can, Not As I Say: Grounding Language in Robotic Affordances. In 6th Annual Conference on Robot Learning.
  2. Training Verifiers to Solve Math Word Problems. ArXiv:2110.14168 [cs], arXiv:2110.14168.
  3. Active Prompting with Chain-of-Thought for Large Language Models. ArXiv:2302.12246 [cs].
  4. Interactive Task Learning: Agents, Robots, and Humans Acquiring New Tasks through Natural Interactions, volume 26 of Strüngmann Forum Reports. Cambridge, MA: MIT Press.
  5. Inner Monologue: Embodied Reasoning through Planning with Language Models. In 6th Annual Conference on Robot Learning.
  6. Language Models can Solve Computer Tasks. ArXiv:2303.17491 [cs].
  7. Learning General and Efficient Representations of Novel Games Through Interactive Instruction. In Proceedings of the Advances in Cognitive Systems Conference. ISBN 0021-9967.
  8. Learning Hierarchical Symbolic Representations to Support Interactive Task Learning and Knowledge Transfer. In Proceedings of IJCAI 2019, 6095–6102. International Joint Conferences on Artificial Intelligence.
  9. Improving Language Model Prompting in Support of Semi-autonomous Task Learning. In Proceedings of the Advances in Cognitive Systems (ACS) Conference.
  10. Integrating Diverse Knowledge Sources for Online One-shot Learning of Novel Tasks. ArXiv:2208.09554 [cs], arXiv:2208.09554.
  11. Interactive Task Learning. IEEE Int. Sys., 32(4): 6–21.
  12. Few-shot Subgoal Planning with Language Models. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL), 5493–5506. Association for Computational Linguistics.
  13. Mininger, A. 2021. Expanding Task Diversity in Explanation-Based Interactive Task Learning. Ph.D. Thesis, University of Michigan, Ann Arbor.
  14. Learning Goal-Oriented Hierarchical Tasks from Situated Interactive Instruction. In Proceedings of the 28t⁢h𝑡ℎ{}^{th}start_FLOATSUPERSCRIPT italic_t italic_h end_FLOATSUPERSCRIPT AAAI Conference on Artificial Intelligence, volume 2, 113–130. AAAI Press.
  15. Acquiring Grounded Representation of Words with Situated Interactive Instruction. Advances in Cognitive Systems, 2: 113–130.
  16. GPT3-to-plan: Extracting plans from text using GPT-3. In Proceedings of ICAPS FinPlan. ArXiv: 2106.07131 [cs].
  17. OpenAI. 2023. GPT-4 Technical Report. arXiv:2303.08774.
  18. Generative Agents: Interactive Simulacra of Human Behavior. In Proceedings of the 36th Annual ACM Symposium on User Interface Software and Technology, 1–22.
  19. Prompt Programming for Large Language Models: Beyond the Few-Shot Paradigm. In Extended Abstracts of the 2021 CHI Conference on Human Factors in Computing Systems, CHI EA ’21. New York, NY, USA: ACM. ISBN 9781450380959.
  20. Richards, T. B. 2023. Auto-GPT: An Autonomous GPT-4 Experiment.
  21. TIDEE: Tidying Up Novel Rooms using Visuo-Semantic Commonsense Priors. In Computer Vision–ECCV 2022, 480–496. Springer.
  22. ProgPrompt: Generating Situated Robot Task Plans using Large Language Models. arXiv:2209.11302.
  23. Cognitive Architectures for Language Agents. ArXiv:2309.02427 [cs].
  24. On the Planning Abilities of Large Language Models (A Critical Investigation with a Proposed Benchmark). arXiv:2302.06706.
  25. Self-Consistency Improves Chain of Thought Reasoning in Language Models. In The Eleventh International Conference on Learning Representations (ICLR 2023).
  26. Ethical and social risks of harm from Language Models. ArXiv:2112.04359 [cs].
  27. TidyBot: Personalized Robot Assistance with Large Language Models. ArXiv:2305.05658 [cs], arXiv:2305.05658.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. James R. Kirk (8 papers)
  2. Robert E. Wray (9 papers)
  3. Peter Lindes (4 papers)
  4. John E. Laird (15 papers)
Citations (1)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets