Comparing the prediction performance of item response theory and machine learning methods on item responses for educational assessments

被引:1
|
作者
Park, Jung Yeon [1 ,2 ]
Dedja, Klest [3 ]
Pliakos, Konstantinos [3 ]
Kim, Jinho [2 ,4 ,5 ]
Joo, Sean [2 ,6 ]
Cornillie, Frederik [2 ]
Vens, Celine [3 ]
Van den Noortgate, Wim [2 ]
机构
[1] George Mason Univ, Coll Educ & Human Dev, 4400 Univ Dr, Fairfax, VA 22030 USA
[2] Katholieke Univ Leuven, Fac Psychol & Educ Sci & Itec, Imec Res Grp, Campus KULAK,Etienne Sabbelaan 51, B-8500 Kortrijk, Belgium
[3] Katholieke Univ Leuven, Dept Publ Hlth & Primary Care & Itec, Imec Res Grp, Campus KULAK,Etienne Sabbelaan 51, B-8500 Kortrijk, Belgium
[4] Univ Seoul, Grad Sch Educ, 163 Seoulsiripdaero, Seoul 02504, South Korea
[5] Univ Seoul, Urban Bigdata AI Inst, 163 Seoulsiripdaero, Seoul 02504, South Korea
[6] Univ Kansas, Dept Educ Psychol, 1450 Jayhawk Blvd, Lawrence, KS 66045 USA
关键词
Item response theory; Explanatory item response model; Machine learning; Background information; Prediction performance; Educational assessment; CLASSIFIERS;
D O I
10.3758/s13428-022-01910-8
中图分类号
B841 [心理学研究方法];
学科分类号
040201 ;
摘要
To obtain more accurate and robust feedback information from the students' assessment outcomes and to communicate it to students and optimize teaching and learning strategies, educational researchers and practitioners must critically reflect on whether the existing methods of data analytics are capable of retrieving the information provided in the database. This study compared and contrasted the prediction performance of an item response theory method, particularly the use of an explanatory item response model (EIRM), and six supervised machine learning (ML) methods for predicting students' item responses in educational assessments, considering student- and item-related background information. Each of seven prediction methods was evaluated through cross-validation approaches under three prediction scenarios: (a) unrealized responses of new students to existing items, (b) unrealized responses of existing students to new items, and (c) missing responses of existing students to existing items. The results of a simulation study and two real-life assessment data examples showed that employing student- and item-related background information in addition to the item response data substantially increases the prediction accuracy for new students or items. We also found that the EIRM is as competitive as the best performing ML methods in predicting the student performance outcomes for the educational assessment datasets.
引用
收藏
页码:2109 / 2124
页数:16
相关论文
共 50 条
  • [21] Item Difficulty Prediction Using Item Text Features: Comparison of Predictive Performance across Machine-Learning Algorithms
    Stepanek, Lubomir
    Dlouha, Jana
    Martinkova, Patricia
    MATHEMATICS, 2023, 11 (19)
  • [22] Item response theory as a feature selection and interpretation tool in the context of machine learning
    Adrienne S. Kline
    Theresa J. B. Kline
    Joon Lee
    Medical & Biological Engineering & Computing, 2021, 59 : 471 - 482
  • [23] Item response theory in high-stakes pharmacy assessments
    Wolcott, Michael D.
    Olsen, Amanda A.
    Augustine, Jill M.
    CURRENTS IN PHARMACY TEACHING AND LEARNING, 2022, 14 (09) : 1206 - 1214
  • [24] Item response theory as a feature selection and interpretation tool in the context of machine learning
    Kline, Adrienne S.
    Kline, Theresa J. B.
    Lee, Joon
    MEDICAL & BIOLOGICAL ENGINEERING & COMPUTING, 2021, 59 (02) : 471 - 482
  • [25] Item response theory in AI: Analysing machine learning classifiers at the instance level
    Martinez-Plumed, Fernando
    Prudencio, Ricardo B. C.
    Martinez-Uso, Adolfo
    Hernandez-Orallo, Jose
    ARTIFICIAL INTELLIGENCE, 2019, 271 : 18 - 42
  • [26] Scalable Learning of Item Response Theory Models
    Frick, Susanne
    Krivosija, Amer
    Munteanu, Alexander
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 238, 2024, 238
  • [27] Measuring student learning with item response theory
    Lee, Young-Jin
    Palazzo, David J.
    Warnakulasooriya, Rasil
    Pritchard, David E.
    PHYSICAL REVIEW SPECIAL TOPICS-PHYSICS EDUCATION RESEARCH, 2008, 4 (01):
  • [28] Assessment of fit of item response theory models used in large-scale educational survey assessments
    van Rijn, Peter W.
    Sinharay, Sandip
    Haberman, Shelby J.
    Johnson, Matthew S.
    LARGE-SCALE ASSESSMENTS IN EDUCATION, 2016, 4 (01)
  • [29] Missing item responses in latent growth analysis: Item response theory versus classical test theory
    Gorter, R.
    Fox, J-P
    Eekhout, I
    Heymans, M. W.
    Twisk, J. W. R.
    STATISTICAL METHODS IN MEDICAL RESEARCH, 2020, 29 (04) : 996 - 1014
  • [30] Semi-Parametric Item Response Theory With O'Sullivan Splines for Item Responses and Response Time
    Liu, Chen-Wei
    APPLIED PSYCHOLOGICAL MEASUREMENT, 2025,