Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
GPT-5.1
GPT-5.1 96 tok/s
Gemini 3.0 Pro 48 tok/s Pro
Gemini 2.5 Flash 155 tok/s Pro
Kimi K2 197 tok/s Pro
Claude Sonnet 4.5 36 tok/s Pro
2000 character limit reached

Developing a Tutoring Dialog Dataset to Optimize LLMs for Educational Use (2410.19231v1)

Published 25 Oct 2024 in cs.CL and cs.AI

Abstract: Recent advances in LLMs have shown promise for scalable educational applications, but their use in dialog-based tutoring systems remains challenging due to the need for effective pedagogical strategies and the high costs associated with expert-curated datasets. Our study explores the use of smaller, more affordable LLMs for one-on-one tutoring in the context of solving reading comprehension problems. We developed a synthetic tutoring dialog dataset, evaluated by human teachers, and fine-tuned a smaller LLM using this dataset. Furthermore, we conducted an interactive experiment comparing the performance of the fine-tuned model with a larger model in real-world tutoring scenarios. Our results show that the fine-tuned model performs on par with the larger model but at a lower cost, demonstrating a viable, cost-effective approach for implementing LLM-based tutoring systems in educational settings.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (29)
  1. Challenges and applications of large language models. arXiv preprint arXiv:2307.10169, 2023.
  2. Are large language models all you need for task-oriented dialogue? In Proceedings of the 24th Annual Meeting of the Special Interest Group on Discourse and Dialogue, pages 216–228, 2023.
  3. The wisdom of practice: Lessons learned from the study of highly effective tutors. In Improving academic achievement, pages 135–158. Elsevier, 2002.
  4. Carol S Dweck. Mindset: The new psychology of success. Random house, 2006.
  5. Efficiently exploring the causal role of contextual moderators in behavioral science. Proceedings of the National Academy of Sciences, 120(1):e2216315120, 2023.
  6. The influence of affective teacher–student relationships on students’ school engagement and achievement: A meta-analytic approach. Review of educational research, 81(4):493–529, 2011.
  7. Robert C Pianta. Teacher-student interactions: Measurement, impacts, improvement, and policy. In Teachers, teaching, and reform, pages 75–93. Routledge, 2017.
  8. Carly D Robinson. A framework for motivating teacher-student relationships. Educational Psychology Review, 34(4):2061–2094, 2022.
  9. Are we there yet?-a systematic literature review on chatbots in education. Frontiers in artificial intelligence, 4:654924, 2021.
  10. Opportunities and challenges in neural dialog tutoring. In Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics, pages 2357–2372, 2023a.
  11. The ai teacher test: Measuring the pedagogical ability of blender and gpt-3 in educational dialogues. In Proceedings of the 15th International Conference on Educational Data Mining, page 522, 2022.
  12. Recipes for building an open-domain chatbot. arXiv preprint arXiv:2004.13637, 2020.
  13. Step-by-step remediation of students’ mathematical mistakes. arXiv preprint arXiv:2310.10648, 2023.
  14. The talkmoves dataset: K-12 mathematics lesson transcripts annotated for teacher and student discursive moves. In Proceedings of the Thirteenth Language Resources and Evaluation Conference, pages 4654–4662, 2022.
  15. CIMA: A large open access dialogue dataset for tutoring. In Proceedings of the Fifteenth Workshop on Innovative Use of NLP for Building Educational Applications, pages 52–64, Seattle, WA, USA → Online, July 2020. Association for Computational Linguistics. URL https://www.aclweb.org/anthology/2020.bea-1.5.
  16. Mathdial: A dialogue tutoring dataset with rich pedagogical properties grounded in math reasoning problems. In Findings of the Association for Computational Linguistics: EMNLP 2023, pages 5602–5621, 2023b.
  17. Training verifiers to solve math word problems. arXiv preprint arXiv:2110.14168, 2021.
  18. Places: Prompting language models for social conversation synthesis. In Findings of the Association for Computational Linguistics: EACL 2023, pages 844–868, 2023.
  19. Socialdial: A benchmark for socially-aware dialogue systems. In Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval, pages 2712–2722, 2023.
  20. Soda: Million-scale dialogue distillation with social commonsense contextualization. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 12930–12949, 2023.
  21. Book2dial: Generating teacher-student interactions from textbooks for cost-effective development of educational chatbots. arXiv preprint arXiv:2403.03307, 2024.
  22. Mistral 7b. arXiv preprint arXiv:2310.06825, 2023.
  23. Mixtral of experts. arXiv preprint arXiv:2401.04088, 2024.
  24. A prompt pattern catalog to enhance prompt engineering with chatgpt. arXiv preprint arXiv:2302.11382, 2023.
  25. Edu-convokit: An open-source library for education conversation data. arXiv e-prints, pages arXiv–2402, 2024.
  26. Computationally identifying funneling and focusing questions in classroom discourse. BEA 2022, page 224, 2022.
  27. The ncte transcripts: A dataset of elementary math classroom transcripts. In Proceedings of the 18th Workshop on Innovative Use of NLP for Building Educational Applications (BEA 2023), pages 528–538, 2023.
  28. Qlora: Efficient finetuning of quantized llms. Advances in Neural Information Processing Systems, 36, 2024.
  29. “mistakes help us grow”: Facilitating and evaluating growth mindset supportive language in classrooms. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 8877–8897, 2023.

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.