Deep learning in the marking of medical student short answer question examinations: Student perceptions and pilot accuracy assessment

被引:0
|
作者
Hollis-Sando, L. [1 ]
Pugh, C. [1 ]
Franke, K. [1 ]
Zerner, T. [1 ]
Tan, Y. [1 ,2 ]
Carneiro, G. [1 ]
van den Hengel, A. [1 ]
Symonds, I. [1 ]
Duggan, P. [1 ]
Bacchi, S. [1 ,2 ]
机构
[1] Univ Adelaide, Adelaide, SA, Australia
[2] Royal Adelaide Hosp, Adelaide, SA, Australia
关键词
deep learning; natural language processing; automation; medical education;
D O I
暂无
中图分类号
G40 [教育学];
学科分类号
040101 ; 120403 ;
摘要
Introduction: Machine learning has previously been applied to text analysis. There is limited data regarding the acceptability or accuracy of such applications in medical education. This project examined medical student opinion regarding computer-based marking and evaluated the accuracy of deep learning (DL), a subtype of machine learning, in the scoring of medical short answer questions (SAQs). Methods: Fourth- and fifth-year medical students undertook an anonymised online examination. Prior to the examination, students completed a survey gauging their opinion on computer-based marking. Questions were marked by humans, and then a DL analysis was conducted using convolutional neural networks. In the DL analysis, following preprocessing, data were split into a training dataset (on which models were developed using 10-fold cross-validation) and a test dataset (on which performance analysis was conducted). Results: One hundred and eighty-one students completed the examination (participation rate 59.0%). While students expressed concern regarding the accuracy of computer-based marking, the majority of students agreed that computer marking would be more objective than human marking (67.0%) and reported they would not object to computer-based marking (55.5%). Regarding automated marking of SAQs, for 1-mark questions, there were consistently high classification accuracies (mean accuracy 0.98). For more complex 2-mark and 3-mark SAQs, in which multiclass classification was required, accuracy was lower (mean 0.65 and 0.59, respectively). Conclusions: Medical students may be supportive of computer-based marking due to its objectivity. DL has the potential to provide accurate marking of written questions, however further research into DL marking of medical examinations is required.
引用
收藏
页码:38 / 48
页数:11
相关论文
共 50 条
  • [1] STUDENT THINKING ON SHORT ANSWER EXAMINATIONS
    Llichty, V. E.
    JOURNAL OF EDUCATIONAL RESEARCH, 1949, 43 (01): : 41 - 48
  • [2] Professionalism and clinical short answer question marking with machine learning
    Lam, Antoinette
    Lam, Lydia
    Blacketer, Charlotte
    Parnis, Roger
    Franke, Kyle
    Wagner, Morganne
    Wang, David
    Tan, Yiran
    Oakden-Rayner, Lauren
    Gallagher, Steve
    Perry, Seth W.
    Licinio, Julio
    Symonds, Ian
    Thomas, Josephine
    Duggan, Paul
    Bacchi, Stephen
    INTERNAL MEDICINE JOURNAL, 2022, 52 (07) : 1268 - 1271
  • [3] Using Marking Criteria to Improve Learning: An Evaluation of Student Perceptions
    Koshy, Swapna
    IMSCI '08: 2ND INTERNATIONAL MULTI-CONFERENCE ON SOCIETY, CYBERNETICS AND INFORMATICS, VOL II, PROCEEDINGS, 2008, : 194 - 198
  • [4] Medical student perception of force application: An accuracy assessment and pilot training program
    Becker, Heather
    Duncan, Riley
    Newsome, D'Angeleau
    Zaremski, Kenneth A.
    Beutel, Bryan G.
    JOURNAL OF EDUCATION AND HEALTH PROMOTION, 2024, 13 (01)
  • [5] Educational benefits of student self-marking of short-answer questions
    Mires, GJ
    Ben-David, MF
    Preece, PE
    Smith, B
    MEDICAL TEACHER, 2001, 23 (05) : 462 - 466
  • [6] Online Learning in Medical Student Clerkship: A Survey of Student Perceptions and Future Directions
    Patel, Rina
    Bannister, Susan L.
    Degelman, Erin
    Sharma, Tejeswin
    Beran, Tanya N.
    Lewis, Melanie
    Novak, Chris
    CUREUS JOURNAL OF MEDICAL SCIENCE, 2024, 16 (02)
  • [7] APEx: A collaborative question database for medical student assessment
    Vieira, Joaquim Edson
    Siqueira, Jose Oliveira
    Sergio Panse Silveira, Paulo
    MEDICAL EDUCATION, 2023, 57 (05) : 456 - 457
  • [8] A Complete Automatic Short Answer Assessment System With Student Identification
    Suwanwiwat, Hemmaphan
    Blumenstein, Michael
    Pal, Umapada
    2015 13TH IAPR INTERNATIONAL CONFERENCE ON DOCUMENT ANALYSIS AND RECOGNITION (ICDAR), 2015, : 611 - 615
  • [9] Cultural differences in student learning: Some implications for assessment and marking
    Burton, LJ
    Dorman, M
    AUSTRALIAN JOURNAL OF PSYCHOLOGY, 2004, 56 : 166 - 166
  • [10] A PILOT-STUDY OF MEDICAL-STUDENT ABUSE - STUDENT PERCEPTIONS OF MISTREATMENT AND MISCONDUCT IN MEDICAL-SCHOOL
    SHEEHAN, KH
    SHEEHAN, DV
    WHITE, K
    LEIBOWITZ, A
    BALDWIN, DC
    JAMA-JOURNAL OF THE AMERICAN MEDICAL ASSOCIATION, 1990, 263 (04): : 533 - 537