Automated Analysis of Algorithm Descriptions Quality, Through Large Language Models

被引:0
|
作者
Sterbini, Andrea [1 ]
Temperini, Marco [2 ]
机构
[1] Sapienza Univ Rome, Dept Comp Sci, Rome, Italy
[2] Sapienza Univ Rome, Dept Comp Control & Management Engn, Rome, Italy
来源
GENERATIVE INTELLIGENCE AND INTELLIGENT TUTORING SYSTEMS, PT I, ITS 2024 | 2024年 / 14798卷
关键词
Large Language Models; LLM-based Text Similarity; Peer Assessment; Automated Assessment;
D O I
10.1007/978-3-031-63028-6_20
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper we propose a method to classify the students' textual descriptions of algorithms. This work is based on a wealth of data (programming tasks, related algorithm descriptions, and Peer Assessment data), coming from 6 years of use of the system Q2A, in a "Fundamentals of Computer Programming" course, given at first year in our university's Computer Science curriculum. The descriptions are submitted, as part of the answer to a computer programming task, through Q2A, and are subject to (formative) Peer Assessment. The proposed classification method aims to support the teacher on the analysis of the quite numerous students' descriptions, in ours as well as in other similar systems. We 1) process the students' submissions, by topic automated extraction (BERTopic) and by separate Large Language Models, 2) compute their degree of suitability as "algorithm description", in a scale from BAD to GOOD, and 3) compare the obtained classification with those coming from the teacher's direct assessment (expert: one of the authors), and from the Peer Assessment. The automated classification does correlate with both the expert classification and the grades given by the peers to the "clarity" of the descriptions. This result is encouraging in view of the production of a Q2A subsystem allowing the teacher to analyse the students' submissions guided by an automated classification, and ultimately support fully automated grading.
引用
收藏
页码:258 / 271
页数:14
相关论文
共 50 条
  • [41] CALLM: Enhancing Clinical Interview Analysis Through Data Augmentation With Large Language Models
    Wu, Yuqi
    Mao, Kaining
    Zhang, Yanbo
    Chen, Jie
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2024, 28 (12) : 7531 - 7542
  • [42] Implications of Large Language Models for Clinical Practice: Ethical Analysis Through the Principlism Framework
    Armitage, Richard C.
    JOURNAL OF EVALUATION IN CLINICAL PRACTICE, 2025, 31 (01)
  • [43] Directions Towards Efficient and Automated Data Wrangling with Large Language Models
    Zhang, Zeyu
    Groth, Paul
    Calixto, Iacer
    Schelter, Sebastian
    2024 IEEE 40TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING WORKSHOP, ICDEW, 2024, : 301 - 304
  • [44] Leveraging Large Language Models for Automated Program Repair in Programming Education
    Murali, Pavithra Sripathanallur
    XRDS: Crossroads, 2025, 31 (02): : 58 - 60
  • [45] Large language models streamline automated machine learning for clinical studies
    Arasteh, Soroosh Tayebi
    Han, Tianyu
    Lotfinia, Mahshad
    Kuhl, Christiane
    Kather, Jakob Nikolas
    Truhn, Daniel
    Nebelung, Sven
    NATURE COMMUNICATIONS, 2024, 15 (01)
  • [46] Automated fact-checking of climate claims with large language models
    Leippold, Markus
    Vaghefi, Saeid Ashraf
    Stammbach, Dominik
    Muccione, Veruska
    Bingler, Julia
    Ni, Jingwei
    Senni, Chiara Colesanti
    Wekhof, Tobias
    Schimanski, Tobias
    Gostlow, Glen
    Yu, Tingyu
    Luterbacher, Juerg
    Huggel, Christian
    NPJ CLIMATE ACTION, 2025, 4 (01):
  • [47] Large language models streamline automated machine learning for clinical studies
    Soroosh Tayebi Arasteh
    Tianyu Han
    Mahshad Lotfinia
    Christiane Kuhl
    Jakob Nikolas Kather
    Daniel Truhn
    Sven Nebelung
    Nature Communications, 15
  • [48] PENTESTGPT: Evaluating and Harnessing Large Language Models for Automated Penetration Testing
    Deng, Gelei
    Liu, Yi
    Mayoral-Vilches, Victor
    Liu, Peng
    Li, Yuekang
    Xu, Yuan
    Zhang, Tianwei
    Liu, Yang
    Pinzger, Martin
    Rass, Stefan
    PROCEEDINGS OF THE 33RD USENIX SECURITY SYMPOSIUM, SECURITY 2024, 2024, : 847 - 864
  • [49] Automated Unit Test Improvement using Large Language Models at Meta
    Alshahwan, Nadia
    Chheda, Jubin
    Finogenova, Anastasia
    Gokkaya, Beliz
    Harman, Mark
    Harper, Inna
    Marginean, Alexandru
    Sengupta, Shubho
    Wang, Eddy
    COMPANION PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON THE FOUNDATIONS OF SOFTWARE ENGINEERING, FSE COMPANION 2024, 2024, : 185 - 196
  • [50] Towards automated phenotype definition extraction using large language models
    Ramya Tekumalla
    Juan M. Banda
    Genomics & Informatics, 22 (1)