Towards Automated Multiple Choice Question Generation and Evaluation: Aligning with Bloom's Taxonomy

被引:1
|
作者
Hwang, Kevin [1 ]
Wang, Kenneth [1 ]
Alomair, Maryam [2 ]
Choa, Fow-Sen [2 ]
Chen, Lujie Karen [2 ]
机构
[1] Glenelg High Sch, Glenelg, MD 21737 USA
[2] Univ Maryland Baltimore Cty, Baltimore, MD 21250 USA
关键词
automated question generation; GPT-4; Bloom's taxonomy; large language models; multiple choice question generation; ITEM WRITING FLAWS;
D O I
10.1007/978-3-031-64299-9_35
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multiple Choice Questions (MCQs) are frequently used for educational assessments for their efficiency in grading and providing feedback. However, manually generatingMCQs has some limitations and challenges. This study explores an AI-driven approach to creating and evaluating Bloom's Taxonomy-aligned college-level biology MCQs using a varied number of shots in few-shot prompting with GPT-4. Shots, or examples of correct prompt-response pairs, were sourced from previously published datasets containing educator-approved MCQs labeled with their Bloom's taxonomy and were matched to prompts via a maximal marginal relevance search. To obtain ground truths to compare GPT-4 against, three expert human evaluators with a minimum of 4 years of educational experience annotated a random sample of the generated questions with regards to relevance to the input prompt, classroom usability, and perceived Bloom's Taxonomy level. Furthermore, we explored the feasibility of an AI-driven evaluation approach that can rate question usability using the Item Writing Flaws (IWFs) framework. We conclude that GPT-4 generally shows promise in generating relevant and usable questions. However, more work needs to be done to improve Bloom-level alignment accuracy (accuracy of alignment between GPT-4's target level and the actual level of the generated question). Moreover, we note that a general inverse relationship exists between alignment accuracy and number of shots. On the other hand, no clear trend between shot number and relevance/usability was observed. These findings shed light on automated question generation and assessment, presenting the potential for advancements in AI-driven educational evaluation methods.
引用
收藏
页码:389 / 396
页数:8
相关论文
共 50 条
  • [1] Cognitive Evaluation of Examinees by Dynamic Question Set Generation based on Bloom's Taxonomy
    Dutta, Anjan
    Chatterjee, Punyasha
    Dey, Nilanjan
    Moreno-Ger, Pablo
    Sen, Soumya
    IETE JOURNAL OF RESEARCH, 2024, 70 (03) : 2570 - 2582
  • [2] Question Guru: An Automated Multiple-Choice Question Generation System
    Gilal, Abdul Rehman
    Waqas, Ahmad
    Talpur, Bandeh Ali
    Abro, Rizwan Ali
    Jaafar, Jafreezal
    Amur, Zaira Hassan
    PROCEEDINGS OF THE 2ND INTERNATIONAL CONFERENCE ON EMERGING TECHNOLOGIES AND INTELLIGENT SYSTEMS, ICETIS 2022, VOL 2, 2023, 573 : 501 - 514
  • [3] Indonesian Question Generation Based on Bloom's Taxonomy Using Text Analysis
    Kusuma, Selvia Ferdiana
    Alhamri, Rinanza Zulmy
    Siahaan, Daniel Oranova
    Fatichah, Chastine
    Naufal, Mohammad Farid
    2018 INTERNATIONAL SEMINAR ON INTELLIGENT TECHNOLOGY AND ITS APPLICATIONS (ISITIA 2018), 2018, : 269 - 274
  • [4] Examining Bloom’s Taxonomy in Multiple Choice Questions: Students’ Approach to Questions
    J. K. Stringer
    Sally A. Santen
    Eun Lee
    Meagan Rawls
    Jean Bailey
    Alicia Richards
    Robert A. Perera
    Diane Biskobing
    Medical Science Educator, 2021, 31 : 1311 - 1317
  • [5] Examining Bloom's Taxonomy in Multiple Choice Questions: Students' Approach to Questions
    Stringer, J. K.
    Santen, Sally A.
    Lee, Eun
    Rawls, Meagan
    Bailey, Jean
    Richards, Alicia
    Perera, Robert A.
    Biskobing, Diane
    MEDICAL SCIENCE EDUCATOR, 2021, 31 (04) : 1311 - 1317
  • [6] Crowdsourcing the Evaluation of Multiple-Choice Questions Using Item-Writing Flaws and Bloom's Taxonomy
    Moore, Steven
    Fang, Ellen
    Nguyen, Huy A.
    Stamper, John
    PROCEEDINGS OF THE TENTH ACM CONFERENCE ON LEARNING @ SCALE, L@S 2023, 2023, : 25 - 34
  • [7] Mining Exam Question based on Bloom's Taxonomy
    Tanalol, Siti Hasnah
    Fattah, Salmah
    Sulong, Rina Suryani
    Mamat, Mazlina
    KMICE 2008 - KNOWLEDGE MANAGEMENT INTERNATIONAL CONFERENCE, 2008 - TRANSFERRING, MANAGING AND MAINTAINING KNOWLEDGE FOR NATION CAPACITY DEVELOPMENT, 2008, : 424 - 427
  • [8] Quality education through writing: aligning learning objectives in learning materials and question papers using Bloom's taxonomy
    Chavda, Mrunal
    Patel, Harsh
    Bhatt, Hetav
    QUALITY ASSURANCE IN EDUCATION, 2024, 32 (01) : 96 - 110
  • [9] Incorporation of Bloom's Taxonomy into Multiple-Choice Examination Questions for a Pharmacotherapeutics Course
    Kim, Myo-Kyoung
    Patel, Rajul A.
    Uchizono, James A.
    Beck, Lynn
    AMERICAN JOURNAL OF PHARMACEUTICAL EDUCATION, 2012, 76 (06)
  • [10] An automatic question generation system using rule-based approach in bloom’s taxonomy
    Gnanasekaran D.
    Kothandaraman R.
    Kaliyan K.
    Recent Advances in Computer Science and Communications, 2021, 14 (05): : 1477 - 1487