Toward In-Context Teaching: Adapting Examples to Students' Misconceptions (2405.04495v1)
Abstract: When a teacher provides examples for a student to study, these examples must be informative, enabling a student to progress from their current state toward a target concept or skill. Good teachers must therefore simultaneously infer what students already know and adapt their teaching to students' changing state of knowledge. There is increasing interest in using computational models, particularly LLMs, as pedagogical tools. As students, LLMs in particular have shown a remarkable ability to adapt to new tasks given small numbers of examples. But how effectively can these models adapt as teachers to students of different types? To study this question, we introduce a suite of models and evaluation methods we call AdapT. AdapT has two components: (1) a collection of simulated Bayesian student models that can be used for evaluation of automated teaching methods; (2) a platform for evaluation with human students, to characterize the real-world effectiveness of these methods. We additionally introduce (3) AToM, a new probabilistic model for adaptive teaching that jointly infers students' past beliefs and optimizes for the correctness of future beliefs. In evaluations of simulated students across three learning domains (fraction arithmetic, English morphology, function learning), AToM systematically outperforms LLM-based and standard Bayesian teaching models. In human experiments, both AToM and LLMs outperform non-adaptive random example selection. Our results highlight both the difficulty of the adaptive teaching task and the potential of learned adaptive models for solving it.
- What learning algorithm is in-context learning? investigations with linear models. In The Eleventh International Conference on Learning Representations.
- Bayesian theory of mind: Modeling joint Belief-Desire attribution. Proceedings of the Annual Meeting of the Cognitive Science Society, 33(33).
- UniMorph 4.0: Universal Morphology. In Proceedings of the Thirteenth Language Resources and Evaluation Conference, pages 840–855, Marseille, France. European Language Resources Association.
- A computational model of fraction arithmetic. Psychol. Rev., 124(5):603–625.
- Language models are few-shot learners. In Advances in Neural Information Processing Systems, volume 33, pages 1877–1901. Curran Associates, Inc.
- From ’why?’ to ’wat!’: Explaining perplexing programs by debugging mental models. In PLATEAU Workshop.
- Learning to teach, teaching to learn.
- Albert Corbett. 2001. Cognitive computer tutors: Solving the Two-Sigma problem. In User Modeling 2001, pages 137–147. Springer Berlin Heidelberg.
- Albert T Corbett and John R Anderson. 1994. Knowledge tracing: Modeling the acquisition of procedural knowledge. User Model. User-adapt Interact., 4(4):253–278.
- Legibility and predictability of robot motion. In Proceedings of the 8th ACM/IEEE International Conference on Human-Robot Interaction, HRI ’13, page 301–308. IEEE Press.
- Michael C. Frank and Noah D. Goodman. 2012. Predicting pragmatic reasoning in language games. Science, 336(6084):998–998.
- Cooperative inverse reinforcement learning.
- RK Hambelton and M Jodoin. 2003. Item response theory: models and features.
- Ronald K. Hambleton and Hariharan Swaminathan. 1981. Journal of Educational Measurement, 18(3):178–180.
- Chatgpt for good? on opportunities and challenges of large language models for education. Learning and Individual Differences, 103:102274.
- Eliciting human preferences with language models.
- Smitha Milli and Anca D Dragan. 2019. Literal or pedagogic human? analyzing human model misspecification in objective learning.
- Rethinking the role of demonstrations: What makes in-context learning work? In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 11048–11064, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
- Deep knowledge tracing. CoRR, abs/1506.05908.
- Faster teaching via POMDP planning. Cogn. Sci., 40(6):1290–1332.
- Inferring learners’ knowledge from their actions. Cogn. Sci., 39(3):584–618.
- Can language models teach weaker agents? teacher explanations improve students via personalization.
- Neural theory-of-mind? on the limits of social intelligence in large lms.
- Minding language models’ (lack of) theory of mind: A plug-and-play multi-character belief tracker.
- A rational account of pedagogical reasoning: teaching by, and learning from, examples. Cogn. Psychol., 71:55–89.
- Megha Srivastava and Noah Goodman. 2021. Question generation for adaptive education. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 2: Short Papers), pages 692–701, Online. Association for Computational Linguistics.
- Extending rational models of communication from beliefs to actions.
- Joshua Tenenbaum. 1999. Rules and similarity in concept learning. In Advances in Neural Information Processing Systems, volume 12. MIT Press.
- Efficient pragmatic program synthesis with informative specifications.
- Bridging the novice-expert gap via models of decision-making: A case study on remediating math mistakes.
- Individualized bayesian knowledge tracing models. In Lecture Notes in Computer Science, Lecture notes in computer science, pages 171–180. Springer Berlin Heidelberg, Berlin, Heidelberg.
- An overview of machine teaching.
- Alexis Ross (13 papers)
- Jacob Andreas (116 papers)