Deep learning in the marking of medical student short answer question examinations: Student perceptions and pilot accuracy assessment

被引:0
|
作者
Hollis-Sando, L. [1 ]
Pugh, C. [1 ]
Franke, K. [1 ]
Zerner, T. [1 ]
Tan, Y. [1 ,2 ]
Carneiro, G. [1 ]
van den Hengel, A. [1 ]
Symonds, I. [1 ]
Duggan, P. [1 ]
Bacchi, S. [1 ,2 ]
机构
[1] Univ Adelaide, Adelaide, SA, Australia
[2] Royal Adelaide Hosp, Adelaide, SA, Australia
关键词
deep learning; natural language processing; automation; medical education;
D O I
暂无
中图分类号
G40 [教育学];
学科分类号
040101 ; 120403 ;
摘要
Introduction: Machine learning has previously been applied to text analysis. There is limited data regarding the acceptability or accuracy of such applications in medical education. This project examined medical student opinion regarding computer-based marking and evaluated the accuracy of deep learning (DL), a subtype of machine learning, in the scoring of medical short answer questions (SAQs). Methods: Fourth- and fifth-year medical students undertook an anonymised online examination. Prior to the examination, students completed a survey gauging their opinion on computer-based marking. Questions were marked by humans, and then a DL analysis was conducted using convolutional neural networks. In the DL analysis, following preprocessing, data were split into a training dataset (on which models were developed using 10-fold cross-validation) and a test dataset (on which performance analysis was conducted). Results: One hundred and eighty-one students completed the examination (participation rate 59.0%). While students expressed concern regarding the accuracy of computer-based marking, the majority of students agreed that computer marking would be more objective than human marking (67.0%) and reported they would not object to computer-based marking (55.5%). Regarding automated marking of SAQs, for 1-mark questions, there were consistently high classification accuracies (mean accuracy 0.98). For more complex 2-mark and 3-mark SAQs, in which multiclass classification was required, accuracy was lower (mean 0.65 and 0.59, respectively). Conclusions: Medical students may be supportive of computer-based marking due to its objectivity. DL has the potential to provide accurate marking of written questions, however further research into DL marking of medical examinations is required.
引用
收藏
页码:38 / 48
页数:11
相关论文
共 50 条
  • [41] Understanding Medical Student Disengagement from Small Group Learning Activities by Comparing Learning Activities by Comparing Faculty and Student Perceptions
    Reji, Merin
    Harris, David M.
    FASEB JOURNAL, 2019, 33
  • [42] Off-line Restricted-Set Handwritten Word Recognition for Student Identification in a Short Answer Question Automated Assessment System
    Suwanwiwat, Hemmaphan
    Vu Nguyen
    Blumenstein, Michael
    2012 12TH INTERNATIONAL CONFERENCE ON HYBRID INTELLIGENT SYSTEMS (HIS), 2012, : 167 - 172
  • [43] Workshop for Basic Gynaecological Examinations: Improving Medical Student Learning through Clinical Simulation
    Cunarro-Lopez, Yolanda
    Llanos, Lucia Sanchez
    Hernandez, Ignacio Cueto
    De Zumarraga, Blanca Gonzalez-Garzon
    Recarte, Maria Del Pilar Pintado
    Labarta, Francisco Javier Ruiz
    Cano-Valderrama, Oscar
    Ocana, Olga Aedo
    Lucas, Raquel Perez
    Beneitez, Maria Del Carmen Vinuela
    Angullo, Zurine Raquel Reyes
    Munoz, Maria Fernandez
    Moreno, Juan Manuel Pina
    Rodriguez, Maria Mercedes Sanchez
    Rodriguez, Rocio Aracil
    Burrel, Laura Perez
    Prat, Ainoa Saez
    Lopez, Andrea Fraile
    Del Rio, Beatriz Gutierrez
    Coronado, Maria de Guadalupe Quintana
    Cisternas, Tamara
    Hidalgo, Marta Feltrer
    Navarro, Pablo Gonzalez
    Ortega, Miguel A.
    Ramon y Cajal, Carlos Lopez
    De Leon-Luis, Juan Antonio
    HEALTHCARE, 2023, 11 (16)
  • [44] Answering the Question: Student Perceptions of Personalized Education and the Construct's Relationship to Learning Outcomes
    Waldeck, Jennifer
    COMMUNICATION EDUCATION, 2007, 56 (04) : 409 - 432
  • [45] Student perceptions and learning of the engineering design process: an assessment at the freshmen level
    Schubert, Thomas F., Jr.
    Jacobitz, Frank G.
    Kim, Ernest M.
    RESEARCH IN ENGINEERING DESIGN, 2012, 23 (03) : 177 - 190
  • [46] Assessment to Optimize Learning Strategies: a Qualitative Study of Student and Faculty Perceptions
    Wlodarczyk, Susan
    Muller-Juge, Virginie
    Hauer, Karen E.
    Tong, Michelle S.
    Ransohoff, Amy
    Boscardin, Christy
    TEACHING AND LEARNING IN MEDICINE, 2021, 33 (03) : 245 - 257
  • [47] Assessment of Student Performance and Perceptions of Learning in a Medication Use Evaluation Project
    Beckett, Robert D.
    Ersin, Oezlem H.
    Chen, Yu-Chieh
    AMERICAN JOURNAL OF PHARMACEUTICAL EDUCATION, 2014, 78 (08)
  • [48] Student perceptions and learning of the engineering design process: an assessment at the freshmen level
    Thomas F. Schubert
    Frank G. Jacobitz
    Ernest M. Kim
    Research in Engineering Design, 2012, 23 : 177 - 190
  • [49] A Pilot Study Comparing Cued Versus Recognition Recall Question Design on Medical Student Utilization, Effectiveness, and Perceptions of Pharmacology Educational Games
    Jason Lee
    Melissa Cowan
    Laurel Gorman
    Medical Science Educator, 2019, 29 : 901 - 904
  • [50] A Pilot Study Comparing Cued Versus Recognition Recall Question Design on Medical Student Utilization, Effectiveness, and Perceptions of Pharmacology Educational Games
    Lee, Jason
    Cowan, Melissa
    Gorman, Laurel
    MEDICAL SCIENCE EDUCATOR, 2019, 29 (04) : 901 - 904