Automated Assessment of Students' Code Comprehension using LLMs

被引:0
|
作者
Oli, Priti [1 ]
Banjade, Rabin [1 ]
Chapagain, Jeevan [1 ]
Rus, Vasile [1 ]
机构
[1] Univ Memphis, Memphis, TN 38152 USA
来源
AI FOR EDUCATION WORKSHOP | 2024年 / 257卷
关键词
Automated Assessment; Large Language Model; Code Comprehension; Self-Explanation;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Assessing students' answers, particularly natural language answers, is a crucial challenge in the field of education. Advances in transformer-based models such as Large Language Models (LLMs), have led to significant progress in various natural language tasks. Nevertheless, amidst the growing trend of evaluating LLMs across diverse tasks, evaluating LLMs in the realm of automated answer assessment has not received much attention. To address this gap, we explore the potential of using LLMs for automated assessment of student's short and open-ended answers in program comprehension tasks. Particularly, we use LLMs to compare students' explanations with expert explanations in the context of line-by-line explanations of computer programs. For comparison purposes, we assess both decoder-only Large Language Models (LLMs) and encoder-based Semantic Textual Similarity (STS) models in the context of assessing the correctness of students' explanation of computer code. Our findings indicate that decoder-only LLMs, when prompted in few-shot and chain-of-thought setting perform comparable to fine-tuned encoder-based models in evaluating students' short answers in the programming domain.
引用
收藏
页码:118 / 128
页数:11
相关论文
共 50 条
  • [11] Measuring code comprehension effort using code reading pattern
    Mondal, Sayani
    Das, Partha Pratim
    Rudra, Titas Bhattacharjee
    SADHANA-ACADEMY PROCEEDINGS IN ENGINEERING SCIENCES, 2022, 47 (03):
  • [12] Code Confabulator: Harnessing LLMs to Compile Code for Visualization
    Amrita School of Computing, Amrita Vishwa Vidyapeetham, Department of Computer Science and Engineering, Bengaluru, India
    Int. Conf. Comput. Commun. Netw. Technol., ICCCNT,
  • [13] LLMs: Understanding Code Syntax and Semantics for Code Analysis
    Ma, Wei
    Wang, Wenhan
    Liu, Ye
    Liu, Shangqing
    Hu, Qiang
    Li, Li
    Liu, Yang
    Lin, Zhihao
    Zhang, Cen
    Nie, Liming
    arXiv, 2023,
  • [14] Evaluating Students' Programs Using Automated Assessment - A Case Study
    English, John
    Rosenthal, Tammy
    ITICSE 2009: PROCEEDING OF THE 2009 ACM SIGSE ANNUAL CONFERENCE ON INNOVATION AND TECHNOLOGY IN COMPUTER SCIENCE EDUCATION, 2009, : 371 - 371
  • [15] Poster: Benchmarking of Code Generative LLMs
    Rahman, Mirza Masfiqur
    Kundu, Ashish
    Bertino, Elisa
    2024 IEEE 44TH INTERNATIONAL CONFERENCE ON DISTRIBUTED COMPUTING SYSTEMS, ICDCS 2024, 2024, : 1448 - 1449
  • [16] LLMs for Code: The Potential, Prospects, and Problems
    Sharma, Tushar
    IEEE 21ST INTERNATIONAL CONFERENCE ON SOFTWARE ARCHITECTURE COMPANION, ICSA-C 2024, 2024, : 373 - 374
  • [17] COCONUT: COde COmprehension Nurturant Using Traceability
    De Lucia, Andrea
    Di Penta, Massimiliano
    Oliveto, Rocco
    Zurolo, Francesco
    ICSM 2006: 22ND IEEE INTERNATIONAL CONFERENCE ON SOFTWARE MAINTENANCE, PROCEEDINGS, 2006, : 274 - +
  • [18] Code Needs Comments: Enhancing Code LLMs with Comment Augmentation
    Song, Demin
    Guo, Honglin
    Zhou, Yunhua
    Xing, Shuhao
    Wang, Yudong
    Song, Zifan
    Zhang, Wenwei
    Guo, Qipeng
    Yan, Hang
    Qiu, Xipeng
    Lin, Dahua
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 13640 - 13656
  • [19] Disaster Damage Assessment Using LLMs and Image Analysis
    Tani, Masahiro
    Terao, Makoto
    Sogi, Naoya
    Shibata, Takashi
    Senzaki, Kenta
    Rodrigues, Royston
    NEC Technical Journal, 2024, 17 (02): : 51 - 53
  • [20] Concept Induction Using LLMs: A User Experiment for Assessment
    Barua, Adrita
    Widmer, Cara
    Hitzler, Pascal
    NEURAL-SYMBOLIC LEARNING AND REASONING, PT II, NESY 2024, 2024, 14980 : 132 - 148