Generative Artificial Intelligence for Chest Radiograph Interpretation in the Emergency Department

被引:32
|
作者
Huang, Jonathan [1 ,3 ,4 ]
Neill, Luke [1 ]
Wittbrodt, Matthew [2 ]
Melnick, David [2 ]
Klug, Matthew [2 ]
Thompson, Michael [2 ]
Bailitz, John [1 ]
Loftus, Timothy [1 ]
Malik, Sanjeev [1 ]
Phull, Amit [1 ]
Weston, Victoria [1 ]
Heller, J. Alex [2 ]
Etemadi, Mozziyar [2 ,3 ,4 ,5 ]
机构
[1] Northwestern Univ, Feinberg Sch Med, Dept Emergency Med, Chicago, IL USA
[2] Northwestern Med Informat Serv, Res & Dev, Chicago, IL USA
[3] Northwestern Univ, Feinberg Sch Med, Dept Anesthesiol, Chicago, IL USA
[4] Northwestern Univ, McCormick Sch Engn, Dept Biomed Engn, Evanston, IL USA
[5] Northwestern Med, Adv Technol, Chicago, IL 60611 USA
关键词
ACCURACY;
D O I
10.1001/jamanetworkopen.2023.36100
中图分类号
R5 [内科学];
学科分类号
1002 ; 100201 ;
摘要
Importance Multimodal generative artificial intelligence (AI) methodologies have the potential to optimize emergency department care by producing draft radiology reports from input images.Objective To evaluate the accuracy and quality of AI-generated chest radiograph interpretations in the emergency department setting.Design, Setting, and Participants This was a retrospective diagnostic study of 500 randomly sampled emergency department encounters at a tertiary care institution including chest radiographs interpreted by both a teleradiology service and on-site attending radiologist from January 2022 to January 2023. An AI interpretation was generated for each radiograph. The 3 radiograph interpretations were each rated in duplicate by 6 emergency department physicians using a 5-point Likert scale.Main Outcomes and Measures The primary outcome was any difference in Likert scores between radiologist, AI, and teleradiology reports, using a cumulative link mixed model. Secondary analyses compared the probability of each report type containing no clinically significant discrepancy with further stratification by finding presence, using a logistic mixed-effects model. Physician comments on discrepancies were recorded.Results A total of 500 ED studies were included from 500 unique patients with a mean (SD) age of 53.3 (21.6) years; 282 patients (56.4%) were female. There was a significant association of report type with ratings, with post hoc tests revealing significantly greater scores for AI (mean [SE] score, 3.22 [0.34]; P < .001) and radiologist (mean [SE] score, 3.34 [0.34]; P < .001) reports compared with teleradiology (mean [SE] score, 2.74 [0.34]) reports. AI and radiologist reports were not significantly different. On secondary analysis, there was no difference in the probability of no clinically significant discrepancy between the 3 report types. Further stratification of reports by presence of cardiomegaly, pulmonary edema, pleural effusion, infiltrate, pneumothorax, and support devices also yielded no difference in the probability of containing no clinically significant discrepancy between the report types.Conclusions and Relevance In a representative sample of emergency department chest radiographs, results suggest that the generative AI model produced reports of similar clinical accuracy and textual quality to radiologist reports while providing higher textual quality than teleradiologist reports. Implementation of the model in the clinical workflow could enable timely alerts to life-threatening pathology while aiding imaging interpretation and documentation.
引用
收藏
页数:10
相关论文
共 50 条
  • [41] Electronic radiograph reports in the emergency department
    Buckley, A.
    Eccles, S.
    Gant, L.
    ANNALS OF EMERGENCY MEDICINE, 2008, 51 (04) : 498 - 498
  • [42] Facilitating Radiograph Interpretation: Refined Generative Models for Precise Bone Suppression in Chest X-rays
    Ibrahim, Samar
    Selim, Sahar
    Elattar, Mustafa
    JOURNAL OF IMAGING INFORMATICS IN MEDICINE, 2025,
  • [43] Documentation of chest radiograph interpretation in the acute setting
    Elswood, T.
    Routledge, B.
    SCOTTISH MEDICAL JOURNAL, 2013, 58 (01) : E59 - E60
  • [44] Impact of Clinical History on Chest Radiograph Interpretation
    Test, Matthew
    Shah, Samir S.
    Monuteaux, Michael
    Ambroggio, Lilliam
    Lee, Edward Y.
    Markowitz, Richard I.
    Bixby, Sarah
    Diperna, Stephanie
    Servaes, Sabah
    Hellinger, Jeffrey C.
    Neuman, Mark I.
    JOURNAL OF HOSPITAL MEDICINE, 2013, 8 (07) : 359 - 364
  • [45] Chest radiograph interpretation skills among anesthesiologists
    Dhar, P
    Kaufman, B
    Fermon, C
    O'Neill, D
    Sutin, K
    Wahlander, S
    Turndorf, H
    ANESTHESIOLOGY, 2000, : U14 - U14
  • [46] The Added Value of the Lateral Chest Radiograph for Diagnosing Community Acquired Pneumonia in the Pediatric Emergency Department
    Soudack, Michelle
    Plotkin, Semion
    Ben-Shlush, Aviva
    Raviv-Zilka, Lisa
    Jacobson, Jeffrey M.
    Benacon, Michael
    Augarten, Arie
    ISRAEL MEDICAL ASSOCIATION JOURNAL, 2018, 20 (01): : 5 - 8
  • [47] THE INFLUENCE OF THE LORDOTIC PROJECTION ON THE INTERPRETATION OF THE CHEST RADIOGRAPH
    HOLLMAN, AS
    ADAMS, FG
    CLINICAL RADIOLOGY, 1989, 40 (04) : 360 - 364
  • [48] Generative Artificial Intelligence
    Lee, Christoph I.
    Chen, Jonathan H.
    Kohli, Marc D.
    Smith, Andrew D.
    Liao, Joshua M.
    JOURNAL OF THE AMERICAN COLLEGE OF RADIOLOGY, 2024, 21 (08) : 1318 - 1320
  • [49] Generative Artificial Intelligence
    Hawk, Heather
    Coriasco, Michael
    Jones, Jeffrey R.
    NURSE EDUCATOR, 2025, 50 (01) : 18 - 22
  • [50] Generative artificial intelligence
    Banh, Leonardo
    Strobel, Gero
    ELECTRONIC MARKETS, 2023, 33 (01)