Classification of mathematical test questions using machine learning on datasets of learning management system questions

被引:4
|
作者
Kim, Gun Il [1 ]
Kim, Sungtae [2 ]
Jang, Beakcheol [1 ]
机构
[1] Yonsei Univ, Grad Sch Informat, Seoul, South Korea
[2] ABLE EduTech Inc, Seoul, South Korea
来源
PLOS ONE | 2023年 / 18卷 / 10期
关键词
D O I
10.1371/journal.pone.0286989
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Every student has a varied level of mathematical proficiency. Therefore, it is important to provide them with questions accordingly. Owing to advances in technology and artificial intelligence, the Learning Management System (LMS) has become a popular application to conduct online learning for students. The LMS can store multiple pieces of information on students through an online database, enabling it to recommend appropriate questions for each student based on an analysis of their previous responses to questions. Particularly, the LMS manages learners and provides an online platform that can evaluate their skills. Questions need to be classified according to their difficulty level so that the LMS can recommend them to learners appropriately and thereby increase their learning efficiency. In this study, we classified large-scale mathematical test items provided by ABLE Tech, which supports LMS-based online mathematical education platforms, using various machine learning techniques according to the difficulty levels of the questions. First, through t-test analysis, we identified the significant correlation variables according to the difficulty level. The t-test results showed that answer rate, type of question, and solution time were positively correlated with the difficulty of the question. Second, items were classified according to their difficulty level using various machine learning models, such as logistic regression (LR), random forest (RF), and extreme gradient boosting (xgboost). Accuracy, precision, recall, F1 score, the area under the curve of the receiver operating curve (AUC-ROC), Cohen's Kappa and Matthew's correlation coefficient (MCC) scores were used as the evaluation metrics. The correct answer rate, question type, and time for solving a question correlated significantly with the difficulty level. The machine learning-based xgboost model outperformed the statistical machine learning models, with a 85.7% accuracy, and 85.8% F1 score. These results can be used as an auxiliary tool in recommending suitable mathematical questions to various learners based on their difficulty level.
引用
收藏
页数:17
相关论文
共 50 条
  • [31] Classification Comparison of Machine Learning Algorithms Using Two Independent CAD Datasets
    Yuvali, Meliz
    Yaman, Belma
    Tosun, Oezguer
    MATHEMATICS, 2022, 10 (03)
  • [32] Effective Handling of Missing Values in Datasets for Classification Using Machine Learning Methods
    Palanivinayagam, Ashokkumar
    Damasevicius, Robertas
    INFORMATION, 2023, 14 (02)
  • [33] Machine learning for COVID-19-asking the right questions
    Bachtiger, Patrik
    Peters, Nicholas S.
    Walsh, Simon L. F.
    LANCET DIGITAL HEALTH, 2020, 2 (08): : E391 - E392
  • [34] A machine learning approach to answering questions for reading comprehension tests
    Ng, HT
    Teo, LH
    Kwan, JLP
    PROCEEDINGS OF THE 2000 JOINT SIGDAT CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND VERY LARGE CORPORA, 2000, : 124 - 132
  • [35] LEARNING OF EPILEPSY BASED ON QUESTIONS: 120 QUESTIONS AND ANSWERS
    Gomez Zuluaga, Olga
    IATREIA, 2011, 24 (04) : 441 - 441
  • [36] Machine learning in liver transplantation: a tool for some unsolved questions?
    Ferrarese, Alberto
    Sartori, Giuseppe
    Orru, Graziella
    Frigo, Anna Chiara
    Pelizzaro, Filippo
    Burra, Patrizia
    Senzolo, Marco
    TRANSPLANT INTERNATIONAL, 2021, 34 (03) : 398 - 411
  • [37] Annotating Exam Questions Through Automatic Learning Concept Classification
    Begusic, Domagoj
    Pintar, Damir
    Skopljanac-Macina, Frano
    Vranic, Mihaela
    2018 26TH INTERNATIONAL CONFERENCE ON SOFTWARE, TELECOMMUNICATIONS AND COMPUTER NETWORKS (SOFTCOM), 2018, : 123 - 128
  • [38] Frailty Modeling Using Machine Learning Methodologies: A Systematic Review With Discussions on Outstanding Questions
    Yang, Hongfei
    Chang, Jiangeng
    He, Wenbo
    Wee, Caitlin Fern
    Yit, John Soong Tshon
    Feng, Mengling
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2025, 29 (01) : 631 - 642
  • [39] Reading to answer questions: Using student questions to teach across diverse learning styles
    Bansenauer, B
    SELECTED PAPERS FROM THE 7TH NATIONAL CONFERENCE ON COLLEGE TEACHING AND LEARNING, 1996, : 15 - 23
  • [40] A test of fundamental questions in mimicry theory using long-term datasets
    Long, Elizabeth C.
    Edwards, Kyle F.
    Shapiro, Arthur M.
    BIOLOGICAL JOURNAL OF THE LINNEAN SOCIETY, 2015, 116 (03) : 487 - 494