Performance of ChatGPT on the Plastic Surgery Inservice Training Examination

被引:39
|
作者
Gupta, Rohun [1 ,3 ]
Herzog, Isabel [2 ]
Park, John B. [2 ]
Weisberger, Joseph
Firouzbakht, Peter [1 ]
Ocon, Vanessa
Chao, John [2 ]
Lee, Edward S.
Mailey, Brian A.
机构
[1] St Louis Univ, Dept Surg, Div Plast Surg, Sch Med, St Louis, MO USA
[2] Rutgers New Jersey Sch Med, Dept Plast Surg, Newark, NJ USA
[3] SLUCare Acad Pavil, 1008 S Spring Ave,Suite,1500 St Louis, St Louis, MO 63110 USA
关键词
D O I
10.1093/asj/sjad128
中图分类号
R61 [外科手术学];
学科分类号
摘要
Background Developed originally as a tool for resident self-evaluation, the Plastic Surgery Inservice Training Examination (PSITE) has become a standardized tool adopted by Plastic Surgery residency programs. The introduction of large language models (LLMs), such as ChatGPT (OpenAI, San Francisco, CA), has demonstrated the potential to help propel the field of Plastic Surgery. Objectives The authors of this study wanted to assess whether or not ChatGPT could be utilized as a tool in resident education by assessing its accuracy on the PSITE. Methods Questions were obtained from the 2022 PSITE, which was present on the American Council of Academic Plastic Surgeons (ACAPS) website. Questions containing images or tables were carefully inspected and flagged before being inputted into ChatGPT. All responses by ChatGPT were qualified utilizing the properties of natural coherence. Responses that were found to be incorrect were divided into the following categories: logical, informational, or explicit fallacy. Results ChatGPT answered a total of 242 questions with an accuracy of 54.96%. The software incorporated logical reasoning in 88.8% of questions, internal information in 95.5% of questions, and external information in 92.1% of questions. When stratified by correct and incorrect responses, we determined that there was a statistically significant difference in ChatGPT's use of external information (P < .05). Conclusions ChatGPT is a versatile tool that has the potential to impact resident education by providing general knowledge, clarifying information, providing case-based learning, and promoting evidence-based medicine. With advancements in LLM and artificial intelligence (AI), it is possible that ChatGPT may be an impactful tool for resident education within Plastic Surgery.
引用
收藏
页码:NP1078 / NP1082
页数:5
相关论文
共 50 条
  • [1] Commentary on: Performance of ChatGPT on the Plastic Surgery Inservice Training Examination
    Cevallos, Priscila C.
    Nazerali, Rahim S.
    [J]. AESTHETIC SURGERY JOURNAL, 2023, 43 (12) : NP1083 - NP1084
  • [2] Performance of ChatGPT on the Plastic Surgery Inservice Training Examination (May , sjad316, 2023)
    Gupta, Rohun
    Herzog, Isabel
    Park, John
    Weisberger, Joseph
    Firouzbakht, Peter
    Ocon, Vanessa
    Chao, John
    Lee, Edward
    Mailey, Brian
    [J]. AESTHETIC SURGERY JOURNAL, 2024, 44 (02) : 232 - 232
  • [3] GPT-4 and plastic surgery inservice training examination
    Daungsupawong, Hinpetch
    Wiwanitkit, Viroj
    [J]. JOURNAL OF PLASTIC RECONSTRUCTIVE AND AESTHETIC SURGERY, 2024, 88 : 71 - 72
  • [4] Applying GPT-4 to the Plastic Surgery Inservice Training Examination
    Gupta, Rohun
    Park, John B.
    Herzog, Isabel
    Yosufi, Nahid
    Mangan, Amelia
    Firouzbakht, Peter K.
    Mailey, Brian A.
    [J]. JOURNAL OF PLASTIC RECONSTRUCTIVE AND AESTHETIC SURGERY, 2023, 87 : 78 - 82
  • [5] Applying GPT-4 to the plastic surgery inservice training examination
    Zhao, Jiuli
    Du, Hong
    [J]. JOURNAL OF PLASTIC RECONSTRUCTIVE AND AESTHETIC SURGERY, 2024, 91 : 225 - 226
  • [6] Reply to commentary on GPT-4 and plastic surgery inservice training examination
    Gupta, Rohun
    Spence, Lawson E.
    Firouzbakht, Peter K.
    Mailey, Brian A.
    [J]. JOURNAL OF PLASTIC RECONSTRUCTIVE AND AESTHETIC SURGERY, 2024, 88 : 66 - 66
  • [7] Evaluating the performance of ChatGPT-3.5 and ChatGPT-4 on the Taiwan plastic surgery board examination
    Hsieh, Ching-Hua
    Hsieh, Hsiao-Yun
    Lin, Hui-Ping
    [J]. HELIYON, 2024, 10 (14)
  • [8] Performance of ChatGPT on American Board of Surgery In-Training Examination Preparation Questions
    Tran, Catherine G.
    Chang, Jeremy
    Sherman, Scott K.
    De Andrade, James P.
    [J]. JOURNAL OF SURGICAL RESEARCH, 2024, 299 : 329 - 335
  • [9] ChatGPT Is Equivalent to First-Year Plastic Surgery Residents: Evaluation of ChatGPT on the Plastic Surgery In-service Examination
    Humar, Pooja
    Asaad, Malke
    Bengur, Fuat Baris
    Nguyen, Vu
    [J]. AESTHETIC SURGERY JOURNAL, 2023, 43 (12) : NP1085 - NP1089
  • [10] Relationships between Resident Inservice Training Examination and neurology board examination performance
    Juel, VC
    Johnston, KC
    [J]. NEUROLOGY, 2002, 58 (07) : A21 - A22