How Teachers Can Use Large Language Models and Bloom's Taxonomy to Create Educational Quizzes (2401.05914v1)
Abstract: Question generation (QG) is a natural language processing task with an abundance of potential benefits and use cases in the educational domain. In order for this potential to be realized, QG systems must be designed and validated with pedagogical needs in mind. However, little research has assessed or designed QG approaches with the input from real teachers or students. This paper applies a LLM-based QG approach where questions are generated with learning goals derived from Bloom's taxonomy. The automatically generated questions are used in multiple experiments designed to assess how teachers use them in practice. The results demonstrate that teachers prefer to write quizzes with automatically generated questions, and that such quizzes have no loss in quality compared to handwritten versions. Further, several metrics indicate that automatically generated questions can even improve the quality of the quizzes created, showing the promise for large scale use of QG in the classroom setting.
- How Useful Are Educational Questions Generated by Large Language Models? In International Conference on Artificial Intelligence in Education, 536–542. Springer.
- ChatGPT for good? On opportunities and challenges of large language models for education. Learning and Individual Differences, 103: 102274.
- Krathwohl, D. R. 2002. A revision of Bloom’s taxonomy: An overview. Theory into practice, 41(4): 212–218.
- A systematic review of automatic question generation for educational purposes. International Journal of Artificial Intelligence in Education, 30: 121–204.
- Quiz design task: Helping teachers create quizzes with automated question generation. arXiv preprint arXiv:2205.01730.
- An application of hierarchical kappa-type statistics in the assessment of majority agreement among multiple observers. Biometrics, 363–374.
- Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. ACM Computing Surveys, 55(9): 1–35.
- Pedagogical evaluation of automatically generated questions. In Intelligent Tutoring Systems: 12th International Conference, ITS 2014, Honolulu, HI, USA, June 5-9, 2014. Proceedings 12, 294–299. Springer.
- Automatic question generation: a review of methodologies, datasets, evaluation metrics, and applications. Progress in Artificial Intelligence, 12(1): 1–32.
- Evaluating content selection in summarization: The pyramid method. In Proceedings of the human language technology conference of the north american chapter of the association for computational linguistics: Hlt-naacl 2004, 145–152.
- Training language models to follow instructions with human feedback.
- Concerning Kendall’s tau, a nonparametric correlation coefficient. Psychological Bulletin, 53(4): 338.
- On the linguistic and pedagogical quality of automatic question generation via neural machine translation. In Technology-Enhanced Learning for a Free, Safe, and Sustainable World: 16th European Conference on Technology Enhanced Learning, EC-TEL 2021, Bolzano, Italy, September 20-24, 2021, Proceedings 16, 289–294. Springer.
- Evaluating AI-Generated Questions: A Mixed-Methods Analysis Using Question Data and Student Perceptions. In Artificial Intelligence in Education: 23rd International Conference, AIED 2022, Durham, UK, July 27–31, 2022, Proceedings, Part I, 344–353. Springer.
- Towards process-oriented, modular, and versatile question generation that meets educational needs. arXiv preprint arXiv:2205.00355.
- Towards human-like educational question generation with large language models. In International conference on artificial intelligence in education, 153–166. Springer.
- A survey of controllable text generation using transformer-based pre-trained language models. arXiv preprint arXiv:2201.05337.
- Sabina Elkins (5 papers)
- Ekaterina Kochmar (33 papers)
- Jackie C. K. Cheung (11 papers)
- Iulian Serban (6 papers)