Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Comparative Study of AI-Generated (GPT-4) and Human-crafted MCQs in Programming Education (2312.03173v1)

Published 5 Dec 2023 in cs.CY, cs.AI, and cs.CL

Abstract: There is a constant need for educators to develop and maintain effective up-to-date assessments. While there is a growing body of research in computing education on utilizing LLMs in generation and engagement with coding exercises, the use of LLMs for generating programming MCQs has not been extensively explored. We analyzed the capability of GPT-4 to produce multiple-choice questions (MCQs) aligned with specific learning objectives (LOs) from Python programming classes in higher education. Specifically, we developed an LLM-powered (GPT-4) system for generation of MCQs from high-level course context and module-level LOs. We evaluated 651 LLM-generated and 449 human-crafted MCQs aligned to 246 LOs from 6 Python courses. We found that GPT-4 was capable of producing MCQs with clear language, a single correct choice, and high-quality distractors. We also observed that the generated MCQs appeared to be well-aligned with the LOs. Our findings can be leveraged by educators wishing to take advantage of the state-of-the-art generative models to support MCQ authoring efforts.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (15)
  1. Jacob Doughty (1 paper)
  2. Zipiao Wan (2 papers)
  3. Anishka Bompelli (1 paper)
  4. Jubahed Qayum (1 paper)
  5. Taozhi Wang (1 paper)
  6. Juran Zhang (1 paper)
  7. Yujia Zheng (34 papers)
  8. Aidan Doyle (3 papers)
  9. Pragnya Sridhar (3 papers)
  10. Arav Agarwal (10 papers)
  11. Christopher Bogart (15 papers)
  12. Eric Keylor (1 paper)
  13. Can Kultur (2 papers)
  14. Jaromir Savelka (47 papers)
  15. Majd Sakr (14 papers)
Citations (32)