Automated Assessment of Students' Code Comprehension using LLMs

被引:0
|
作者
Oli, Priti [1 ]
Banjade, Rabin [1 ]
Chapagain, Jeevan [1 ]
Rus, Vasile [1 ]
机构
[1] Univ Memphis, Memphis, TN 38152 USA
来源
AI FOR EDUCATION WORKSHOP | 2024年 / 257卷
关键词
Automated Assessment; Large Language Model; Code Comprehension; Self-Explanation;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Assessing students' answers, particularly natural language answers, is a crucial challenge in the field of education. Advances in transformer-based models such as Large Language Models (LLMs), have led to significant progress in various natural language tasks. Nevertheless, amidst the growing trend of evaluating LLMs across diverse tasks, evaluating LLMs in the realm of automated answer assessment has not received much attention. To address this gap, we explore the potential of using LLMs for automated assessment of student's short and open-ended answers in program comprehension tasks. Particularly, we use LLMs to compare students' explanations with expert explanations in the context of line-by-line explanations of computer programs. For comparison purposes, we assess both decoder-only Large Language Models (LLMs) and encoder-based Semantic Textual Similarity (STS) models in the context of assessing the correctness of students' explanation of computer code. Our findings indicate that decoder-only LLMs, when prompted in few-shot and chain-of-thought setting perform comparable to fine-tuned encoder-based models in evaluating students' short answers in the programming domain.
引用
收藏
页码:118 / 128
页数:11
相关论文
共 50 条
  • [1] Automated Assessment of Comprehension Strategies from Self-Explanations Using LLMs
    Nicula, Bogdan
    Dascalu, Mihai
    Arner, Tracy
    Balyan, Renu
    McNamara, Danielle S.
    INFORMATION, 2023, 14 (10)
  • [2] Identifying Gaps in Students' Explanations of Code Using LLMs
    Banjade, Rabin
    Oli, Priti
    Sajib, Mahmudul Islam
    Rus, Vasile
    ARTIFICIAL INTELLIGENCE IN EDUCATION, PT II, AIED 2024, 2024, 14830 : 268 - 275
  • [3] An empirical study on the code naturalness modeling capability for LLMs in automated patch correctness assessment
    Li, Yuning
    Zhong, Wenkang
    Shen, Zongwen
    Li, Chuanyi
    Chen, Xiang
    Ge, Jidong
    Luo, Bin
    AUTOMATED SOFTWARE ENGINEERING, 2025, 32 (02)
  • [4] ZeroLeak: Automated Side-Channel Patching in Source Code Using LLMs
    Tol, M. Caner
    Sunar, Berk
    COMPUTER SECURITY-ESORICS 2024, PT I, 2024, 14982 : 290 - 310
  • [5] Automated Assessment of Reporting Completeness in Orthodontic Research Using LLMs: An Observational Study
    Alharbi, Fahad
    Asiri, Saeed
    APPLIED SCIENCES-BASEL, 2024, 14 (22):
  • [6] Towards automated code parallelization through program comprehension
    Martino, B.Di.
    Iannello, G.
    Program Comprehension, Workshop Proceedings, 1994, : 108 - 115
  • [7] Towards Efficient DataWrangling with LLMs using Code Generation
    Li, Xue
    Dohmen, Till
    PROCEEDINGS OF THE 8TH WORKSHOP ON DATA MANAGEMENT FOR END-TO-END MACHINE LEARNING, DEEM 2024, 2024,
  • [8] Towards a tool for rigorous, automated code comprehension using symbolic execution and semantic analysis
    Stewart, MEM
    29th Annual IEEE/NASA Software Engineering Workshop, Proceedings, 2005, : 89 - 96
  • [9] Software Vulnerability and Functionality Assessment using LLMs
    Jensen, Rasmus Ingemann Tuffveson
    Tawosi, Vali
    Alamir, Salwa
    PROCEEDINGS 2024 ACM/IEEE INTERNATIONAL WORKSHOP ON NL-BASED SOFTWARE ENGINEERING, NLBSE 2024, 2024, : 25 - 28
  • [10] Measuring code comprehension effort using code reading pattern
    Sayani Mondal
    Partha Pratim Das
    Titas Bhattacharjee Rudra
    Sādhanā, 47