Classification of mathematical test questions using machine learning on datasets of learning management system questions

被引:4
|
作者
Kim, Gun Il [1 ]
Kim, Sungtae [2 ]
Jang, Beakcheol [1 ]
机构
[1] Yonsei Univ, Grad Sch Informat, Seoul, South Korea
[2] ABLE EduTech Inc, Seoul, South Korea
来源
PLOS ONE | 2023年 / 18卷 / 10期
关键词
D O I
10.1371/journal.pone.0286989
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Every student has a varied level of mathematical proficiency. Therefore, it is important to provide them with questions accordingly. Owing to advances in technology and artificial intelligence, the Learning Management System (LMS) has become a popular application to conduct online learning for students. The LMS can store multiple pieces of information on students through an online database, enabling it to recommend appropriate questions for each student based on an analysis of their previous responses to questions. Particularly, the LMS manages learners and provides an online platform that can evaluate their skills. Questions need to be classified according to their difficulty level so that the LMS can recommend them to learners appropriately and thereby increase their learning efficiency. In this study, we classified large-scale mathematical test items provided by ABLE Tech, which supports LMS-based online mathematical education platforms, using various machine learning techniques according to the difficulty levels of the questions. First, through t-test analysis, we identified the significant correlation variables according to the difficulty level. The t-test results showed that answer rate, type of question, and solution time were positively correlated with the difficulty of the question. Second, items were classified according to their difficulty level using various machine learning models, such as logistic regression (LR), random forest (RF), and extreme gradient boosting (xgboost). Accuracy, precision, recall, F1 score, the area under the curve of the receiver operating curve (AUC-ROC), Cohen's Kappa and Matthew's correlation coefficient (MCC) scores were used as the evaluation metrics. The correct answer rate, question type, and time for solving a question correlated significantly with the difficulty level. The machine learning-based xgboost model outperformed the statistical machine learning models, with a 85.7% accuracy, and 85.8% F1 score. These results can be used as an auxiliary tool in recommending suitable mathematical questions to various learners based on their difficulty level.
引用
收藏
页数:17
相关论文
共 50 条
  • [1] Research of Test Questions Classification Based on Hybrid Frame Mixing Semantic Comprehension and Machine Learning
    Wang, Rihong
    Cui, Xingmei
    Wang, Chenglong
    PROCEEDINGS OF 2017 INTERNATIONAL CONFERENCE ON ROBOTICS AND ARTIFICIAL INTELLIGENCE (ICRAI 2017), 2015, : 1 - 5
  • [2] Autophagy and machine learning: Unanswered questions
    Yang, Ying
    Pan, Zhaoying
    Sun, Jianhui
    Welch, Joshua
    Klionsky, Daniel J.
    BIOCHIMICA ET BIOPHYSICA ACTA-MOLECULAR BASIS OF DISEASE, 2024, 1870 (06):
  • [3] Using machine learning techniques to interpret WH-questions
    Zukerman, I
    Horvitz, E
    39TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, PROCEEDINGS OF THE CONFERENCE, 2001, : 547 - 554
  • [4] Using paired questions to test the effectiveness of active learning.
    Hartshorn, LG
    ABSTRACTS OF PAPERS OF THE AMERICAN CHEMICAL SOCIETY, 2004, 227 : U408 - U408
  • [5] CLASSIFICATION OF STUDENTS' ACHIEVEMENT VIA MACHINE LEARNING BY USING SYSTEM LOGS IN LEARNING MANAGEMENT SYSTEM
    Koyuncu, Ilhan
    Kilic, Abdullah Faruk
    Orhan Goksun, Derya
    TURKISH ONLINE JOURNAL OF DISTANCE EDUCATION, 2022, 23 (03): : 18 - 30
  • [6] Machine Learning and the Police: Asking the Right Questions
    Vestby, Annette
    Vestby, Jonas
    POLICING-A JOURNAL OF POLICY AND PRACTICE, 2021, 15 (01) : 44 - 58
  • [7] Combining Logic and Machine Learning for Answering Questions
    Gloeckner, Ingo
    Pelzer, Bjoern
    EVALUATING SYSTEMS FOR MULTILINGUAL AND MULTIMODAL INFORMATION ACCESS, 2009, 5706 : 401 - +
  • [8] Analyzing the cognitive level of classroom questions using machine learning techniques
    Yahya, Anwar Ali
    Osman, Addin
    Taleb, Ahmad
    Alattab, Ahmed Abdu
    9TH INTERNATIONAL CONFERENCE ON COGNITIVE SCIENCE, 2013, 97 : 587 - 595
  • [9] Answer Extraction for Definition Questions Using Information Gain and Machine Learning
    Instituto Nacional de Astrofísica Óptica y Electrónica, Universidad de la Sierra Juárez, Mexico
    不详
    72840, Mexico
    IFIP Advances in Information and Communication Technology, 2008, (141-150)
  • [10] Answer extraction for definition questions using information gain and machine learning
    Martinez-Gil, Carmen
    Lopez-Lopez, A.
    ARTIFICIAL INTELLIGENCE IN THEORY AND PRACTICE II, 2008, 276 : 141 - 150