Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

AutoTutor meets Large Language Models: A Language Model Tutor with Rich Pedagogy and Guardrails (2402.09216v3)

Published 14 Feb 2024 in cs.CL and cs.HC

Abstract: LLMs have found several use cases in education, ranging from automatic question generation to essay evaluation. In this paper, we explore the potential of using LLMs to author Intelligent Tutoring Systems. A common pitfall of LLMs is their straying from desired pedagogical strategies such as leaking the answer to the student, and in general, providing no guarantees. We posit that while LLMs with certain guardrails can take the place of subject experts, the overall pedagogical design still needs to be handcrafted for the best learning results. Based on this principle, we create a sample end-to-end tutoring system named MWPTutor, which uses LLMs to fill in the state space of a pre-defined finite state transducer. This approach retains the structure and the pedagogy of traditional tutoring systems that has been developed over the years by learning scientists but brings in additional flexibility of LLM-based approaches. Through a human evaluation study on two datasets based on math word problems, we show that our hybrid approach achieves a better overall tutoring score than an instructed, but otherwise free-form, GPT-4. MWPTutor is completely modular and opens up the scope for the community to improve its performance by improving individual modules or using different teaching strategies that it can follow.

Overview of "Scaling the Authoring of AutoTutors with LLMs"

The paper, "Scaling the Authoring of AutoTutors with LLMs," provides a comprehensive exploration of leveraging LLMs for authoring Intelligent Tutoring Systems (ITSs). The authors recognize the potential of LLMs in educational applications and address the limitations of traditional rule-based ITSs, such as AutoTutor, which often require extensive manual labor and face scalability challenges.

Key Contributions

  1. Hybrid Tutoring System - \MB: The authors introduce a new tutoring system named \MB, which incorporates LLMs to fill the state space of a pre-defined finite state transducer. This approach synergizes the structured pedagogy of traditional systems with the flexibility provided by LLMs.
  2. Maintenance of Pedagogical Integrity: While LLMs are employed to generate content, the overarching pedagogical architecture is crafted by learning experts. This ensures that the educational strategies remain robust and are not degraded by the inherent unpredictability of LLM outputs.
  3. Human Evaluation Study: A comparative evaluation using math word problems demonstrates that \MB achieves higher tutoring scores than a free-form GPT-4, maintaining educational efficacy while being modular and extensible.

Numerical Results

The paper presents quantitative evaluations showing that \MB, utilizing GPT-4 as a backend, achieves a perfect success rate on simpler datasets such as MultiArith. It performs competitively on more complex datasets like MathDial, with notable improvements over baseline GPT-4 performance. The system's modular design facilitates further enhancements by allowing adjustments to individual components and teaching strategies.

Implications for AI and Education

  • Theoretical Implications: This research highlights the potential for hybrid models that combine AI with traditional educational frameworks. By employing LLMs primarily for content generation rather than overall pedagogy, it is possible to harness AI capabilities without compromising on educational integrity.
  • Practical Implications: The use of LLMs in generating tutoring content has the potential to significantly reduce the time and resources needed for developing ITSs. This adaptability can enhance the scalability of educational technologies, allowing for wider implementation across diverse subject areas.
  • Speculation on Future Developments: Future advances in AI could further optimize the integration of LLMs into ITSs, potentially allowing for dynamic pedagogical strategies that adapt in real-time to student needs. As AI models become more transparent and controllable, trust and reliability in educational settings could improve, expanding the scope of these applications.

Conclusion

This paper provides valuable insights into the implementation and evaluation of LLM-based components within Intelligent Tutoring Systems. By maintaining a focus on pedagogical soundness while benefiting from the generative capabilities of LLMs, the authors demonstrate a balanced and effective approach to modern educational challenges. Further research could build on these findings, exploring diverse domains and enhancing the interactive capabilities of such systems. The modular and adaptable nature of \MB provides a promising foundation for future developments in AI-driven education.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (33)
  1. Intelligent Tutoring Systems. Science 228, 4698 (1985), 456–462. https://doi.org/10.1126/science.228.4698.456 arXiv:https://www.science.org/doi/pdf/10.1126/science.228.4698.456
  2. David Baidoo-Anu and Leticia Owusu Ansah. 2023. Education in the era of generative artificial intelligence (AI): Understanding the potential benefits of ChatGPT in promoting teaching and learning. Journal of AI 7, 1 (2023), 52–62. https://dergipark.org.tr/en/pub/jai/issue/77844/1337500
  3. Benjamin S Bloom. 1984. The 2 sigma problem: The search for methods of group instruction as effective as one-to-one tutoring. Educational researcher 13, 6 (1984), 4–16.
  4. On the Opportunities and Risks of Foundation Models. arXiv:2108.07258 [cs.LG]
  5. Authoring Conversational Intelligent Tutoring Systems. In Adaptive Instructional Systems, Robert A. Sottilare and Jessica Schwarz (Eds.). Springer International Publishing, Cham, 593–603.
  6. GPTutor: A ChatGPT-Powered Programming Tool for Code Explanation. In Artificial Intelligence in Education. Posters and Late Breaking Results, Workshops and Tutorials, Industry and Innovation Tracks, Practitioners, Doctoral Consortium and Blue Sky - 24th International Conference, AIED 2023, Tokyo, Japan, July 3-7, 2023, Proceedings (Communications in Computer and Information Science, Vol. 1831), Ning Wang, Genaro Rebolledo-Mendez, Vania Dimitrova, Noboru Matsuda, and Olga C. Santos (Eds.). Springer, 321–327. https://doi.org/10.1007/978-3-031-36336-8_50
  7. Scaling Instruction-Finetuned Language Models. arXiv:2210.11416 [cs.LG]
  8. Training Verifiers to Solve Math Word Problems. arXiv preprint arXiv:2110.14168 (2021).
  9. Holistic Analysis of Hallucination in GPT-4V(ision): Bias and Interference Challenges. arXiv:2311.03287 [cs.LG]
  10. Chain-of-Verification Reduces Hallucination in Large Language Models. arXiv:2309.11495 [cs.CL]
  11. Teaching Tactics in AutoTutor. (01 2000).
  12. Arthur C. Graesser. 2016. Conversations with AutoTutor Help Students Learn. International Journal of Artificial Intelligence in Education 26, 1 (01 Mar 2016), 124–132. https://doi.org/10.1007/s40593-015-0086-4
  13. Intelligent Tutoring Systems with Conversational Dialogue. AI Magazine 22, 4 (Dec. 2001), 39. https://doi.org/10.1609/aimag.v22i4.1591
  14. Opening the Door to Non-programmers: Authoring Intelligent Tutor Behavior by Demonstration. In Intelligent Tutoring Systems, James C. Lester, Rosa Maria Vicari, and Fábio Paraguaçu (Eds.). Springer Berlin Heidelberg, Berlin, Heidelberg, 162–174.
  15. Large Language Models are Zero-Shot Reasoners. arXiv:2205.11916 [cs.CL]
  16. Jailbreaking ChatGPT via Prompt Engineering: An Empirical Study. arXiv:2305.13860 [cs.SE]
  17. MathDial: A Dialogue Tutoring Dataset with Rich Pedagogical Properties Grounded in Math Reasoning Problems. arXiv:2305.14536 [cs.CL]
  18. AutoTutor and Family: A Review of 17 Years of Natural Language Tutoring. International Journal of Artificial Intelligence in Education 24 (12 2014). https://doi.org/10.1007/s40593-014-0029-5
  19. Show Your Work: Scratchpads for Intermediate Computation with Language Models. https://arxiv.org/abs/2112.00114
  20. OpenAI. 2023. GPT-4 Technical Report. arXiv:2303.08774 [cs.CL]
  21. José Paladines and Jaime Ramirez. 2020. A Systematic Literature Review of Intelligent Tutoring Systems With Dialogue in Natural Language. IEEE Access 8 (2020), 164246–164267. https://doi.org/10.1109/ACCESS.2020.3021383
  22. Generative AI for Programming Education: Benchmarking ChatGPT, GPT-4, and Human Tutors. In Proceedings of the 2023 ACM Conference on International Computing Education Research - Volume 2 (Chicago, IL, USA) (ICER ’23). Association for Computing Machinery, New York, NY, USA, 41–42. https://doi.org/10.1145/3568812.3603476
  23. ”Call me Kiran” – ChatGPT as a Tutoring Chatbot in a Computer Science Course. In Proceedings of the 26th International Academic Mindtrek Conference. 83–94. https://dl.acm.org/doi/pdf/10.1145/3616961.3616974
  24. Brian Reiser. 2004. Scaffolding Complex Learning: The Mechanisms of Structuring and Problematizing Student Work. Journal Of The Learning Sciences 13 (01 2004), 273–304. https://doi.org/10.1207/s15327809jls1303_2
  25. Decomposition: A K-8 Computational Thinking Learning Trajectory. In Proceedings of the 2018 ACM Conference on International Computing Education Research (Espoo, Finland) (ICER ’18). Association for Computing Machinery, New York, NY, USA, 124–132. https://doi.org/10.1145/3230977.3230979
  26. Ruffle&Riley: Towards the Automated Induction of Conversational Tutoring Systems. arXiv:2310.01420 [cs.CL]
  27. D. Sleeman and J.S. Brown. 1982. Intelligent Tutoring Systems. Academic Press. https://books.google.ch/books?id=pjqcAAAAMAAJ
  28. Ting Song and Kurt Becker. 2014. Expert vs. novice: Problem decomposition/recomposition in engineering design. In 2014 International Conference on interactive collaborative learning (ICL). IEEE, 181–190.
  29. CLASS: A Design Framework for building Intelligent Tutoring Systems based on Learning Science principles. arXiv:2305.13272 [cs.CL]
  30. Ferhat Yarkin and Justin P. Coon. 2022. Simple Gray Coding and LLR Calculation for MDS Modulation Systems. CoRR abs/2201.08237 (2022). arXiv:2201.08237 https://arxiv.org/abs/2201.08237
  31. MetaMath: Bootstrap Your Own Mathematical Questions for Large Language Models. arXiv:2309.12284 [cs.CL]
  32. How Language Model Hallucinations Can Snowball. arXiv:2305.13534 [cs.CL]
  33. BERTScore: Evaluating Text Generation with BERT. arXiv:1904.09675 [cs.CL]
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Sankalan Pal Chowdhury (9 papers)
  2. Vilém Zouhar (41 papers)
  3. Mrinmaya Sachan (124 papers)
Citations (4)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets