Crowdsourcing the Evaluation of Multiple-Choice Questions Using Item-Writing Flaws and Bloom's Taxonomy

被引:2
|
作者
Moore, Steven [1 ]
Fang, Ellen [1 ]
Nguyen, Huy A. [1 ]
Stamper, John [1 ]
机构
[1] Carnegie Mellon Univ, Human Comp Interact, Pittsburgh, PA 15213 USA
关键词
crowdsourcing; learnersourcing; question evaluation; question quality; question generation; QUALITY;
D O I
10.1145/3573051.3593396
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Multiple-choice questions, which are widely used in educational assessments, have the potential to negatively impact student learning and skew analytics when they contain item-writing flaws. Existing methods for evaluating multiple-choice questions in educational contexts tend to focus primarily on machine readability metrics, such as grammar, syntax, and formatting, without considering the intended use of the questions within course materials and their pedagogical implications. In this study, we present the results of crowdsourcing the evaluation of multiple-choice questions based on 15 common item-writing flaws. Through analysis of 80 crowdsourced evaluations on questions from the domains of calculus and chemistry, we found that crowdworkers were able to accurately evaluate the questions, matching 75% of the expert evaluations across multiple questions. They were able to correctly distinguish between two levels of Bloom's Taxonomy for the calculus questions, but were less accurate for chemistry questions. We discuss how to scale this question evaluation process and the implications it has across other domains. This work demonstrates how crowdworkers can be leveraged in the quality evaluation of educational questions, regardless of prior experience or domain knowledge.
引用
收藏
页码:25 / 34
页数:10
相关论文
共 50 条
  • [21] EFFECTS OF PUBLISHING MULTIPLE-CHOICE QUESTIONS ON THEIR SUBSEQUENT ITEM ANALYSIS
    BANDARANAYAKE, RC
    BUZZARD, AJ
    MARSHALL, VC
    AUSTRALIAN AND NEW ZEALAND JOURNAL OF SURGERY, 1990, 60 (12): : 937 - 941
  • [22] IMPROVEMENT OF MULTIPLE-CHOICE QUESTIONS FOR STUDENT EVALUATION
    KNAUFT, D
    NATIONAL ASSOCIATION OF COLLEGE TEACHERS OF AGRICULTURE JOURNAL, 1983, 27 (03): : 38 - 38
  • [23] AN EVALUATION OF THE AUTOMATICALLY GENERATED MULTIPLE-CHOICE QUESTIONS
    Cubric, M.
    Tosic, M.
    EDULEARN14: 6TH INTERNATIONAL CONFERENCE ON EDUCATION AND NEW LEARNING TECHNOLOGIES, 2014, : 1613 - 1613
  • [24] Quantitatively ranking incorrect responses to multiple-choice questions using item response theory
    Smith, Trevor, I
    Louis, Kyle J.
    Ricci, Bartholomew J.
    Bendjilali, Nasrine
    PHYSICAL REVIEW PHYSICS EDUCATION RESEARCH, 2020, 16 (01):
  • [25] Writing Good Multiple-Choice Questions: A Brief Guide for Radiologists
    DiSantis, David J.
    RADIOGRAPHICS, 2013, 33 (07) : 1865 - 1866
  • [26] WRITING MULTIPLE-CHOICE TEST QUESTIONS - ARE THE FOILS FOILING YOU
    VYDARENY, KH
    BLANE, E
    CALHOUN, JG
    INVESTIGATIVE RADIOLOGY, 1985, 20 (06) : S2 - S2
  • [27] Writing cognitive educational objectives and multiple-choice test questions
    Schultheis, NM
    AMERICAN JOURNAL OF HEALTH-SYSTEM PHARMACY, 1998, 55 (22) : 2397 - 2401
  • [28] Item Analysis of Multiple-Choice Questions in Pharmacology in an Indian Medical School
    Adiga, Manoor Narasimha Sachidananda
    Acharya, Swathi
    Holla, Rajendra
    JOURNAL OF HEALTH AND ALLIED SCIENCES NU, 2021, 11 (03): : 130 - 135
  • [29] Analyzing Multiple-Choice Questions by Model Analysis and Item Response Curves
    Wattanakasiwich, P.
    Ananta, S.
    INTERNATIONAL CONFERENCE ON PHYSICS EDUCATION, 2010, 1263 : 245 - 248
  • [30] ITEM SHELLS - A METHOD FOR WRITING EFFECTIVE MULTIPLE-CHOICE TEST ITEMS
    HALADYNA, TM
    SHINDOLL, RR
    EVALUATION & THE HEALTH PROFESSIONS, 1989, 12 (01) : 97 - 106