A Large Language Model to Detect Negated Expressions in Radiology Reports

被引:0
|
作者
Su, Yvonne [1 ]
Babore, Yonatan B. [1 ]
Kahn Jr, Charles E. [1 ,2 ]
机构
[1] Univ Penn, Perelman Sch Med, Dept Radiol, 3400 Spruce St, Philadelphia, PA 19104 USA
[2] Univ Penn, Inst Biomed Informat, Philadelphia, PA 19104 USA
关键词
Large language models; Negated expression (negex) detection; Named entity recognition; Natural language processing; Radiology reports; SYSTEM;
D O I
10.1007/s10278-024-01274-9
中图分类号
R8 [特种医学]; R445 [影像诊断学];
学科分类号
1002 ; 100207 ; 1009 ;
摘要
Natural language processing (NLP) is crucial to extract information accurately from unstructured text to provide insights for clinical decision-making, quality improvement, and medical research. This study compared the performance of a rule-based NLP system and a medical-domain transformer-based model to detect negated concepts in radiology reports. Using a corpus of 984 de-identified radiology reports from a large U.S.-based academic health system (1000 consecutive reports, excluding 16 duplicates), the investigators compared the rule-based medspaCy system and the Clinical Assertion and Negation Classification Bidirectional Encoder Representations from Transformers (CAN-BERT) system to detect negated expressions of terms from RadLex, the Unified Medical Language System Metathesaurus, and the Radiology Gamuts Ontology. Power analysis determined a sample size of 382 terms to achieve alpha = 0.05 and beta = 0.8 for McNemar's test; based on an estimate of 15% negated terms, 2800 randomly selected terms were annotated manually as negated or not negated. Precision, recall, and F1 of the two models were compared using McNemar's test. Of the 2800 terms, 387 (13.8%) were negated. For negation detection, medspaCy attained a recall of 0.795, precision of 0.356, and F1 of 0.492. CAN-BERT achieved a recall of 0.785, precision of 0.768, and F1 of 0.777. Although recall was not significantly different, CAN-BERT had significantly better precision (chi(2) = 304.64; p < 0.001). The transformer-based CAN-BERT model detected negated terms in radiology reports with high precision and recall; its precision significantly exceeded that of the rule-based medspaCy system. Use of this system will improve data extraction from textual reports to support information retrieval, AI model training, and discovery of causal relationships.
引用
收藏
页数:7
相关论文
共 50 条
  • [1] Generating colloquial radiology reports with large language models
    Tang, Cynthia Crystal
    Nagesh, Supriya
    Fussell, David A.
    Glavis-Bloom, Justin
    Mishra, Nina
    Li, Charles
    Cortes, Gillean
    Hill, Robert
    Zhao, Jasmine
    Gordon, Angellica
    Wright, Joshua
    Troutt, Hayden
    Tarrago, Rod
    Chow, Daniel S.
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2024,
  • [2] Natural Language Processing of Radiology Reports to Detect Complications of Ischemic Stroke
    Miller, Matthew, I
    Orfanoudaki, Agni
    Cronin, Michael
    Saglam, Hanife
    Kim, Ivy So Yeon
    Balogun, Oluwafemi
    Tzalidi, Maria
    Vasilopoulos, Kyriakos
    Fanaropoulou, Georgia
    Fanaropoulou, Nina M.
    Kalin, Jack
    Hutch, Meghan
    Prescott, Brenton R.
    Brush, Benjamin
    Benjamin, Emelia J.
    Shin, Min
    Mian, Asim
    Greer, David M.
    Smirnakis, Stelios M.
    Ong, Charlene J.
    NEUROCRITICAL CARE, 2022, 37 (SUPPL 2) : 291 - 302
  • [3] Natural Language Processing of Radiology Reports to Detect Complications of Ischemic Stroke
    Matthew I. Miller
    Agni Orfanoudaki
    Michael Cronin
    Hanife Saglam
    Ivy So Yeon Kim
    Oluwafemi Balogun
    Maria Tzalidi
    Kyriakos Vasilopoulos
    Georgia Fanaropoulou
    Nina M. Fanaropoulou
    Jack Kalin
    Meghan Hutch
    Brenton R. Prescott
    Benjamin Brush
    Emelia J. Benjamin
    Min Shin
    Asim Mian
    David M. Greer
    Stelios M. Smirnakis
    Charlene J. Ong
    Neurocritical Care, 2022, 37 : 291 - 302
  • [4] A scoping review of large language model based approaches for information extraction from radiology reports
    Reichenpfader, Daniel
    Muller, Henning
    Denecke, Kerstin
    NPJ DIGITAL MEDICINE, 2024, 7 (01):
  • [5] Feasibility of Using the Privacy-preserving Large Language Model Vicuna for Labeling Radiology Reports
    Mukherjee, Pritam
    Hou, Benjamin
    Lanfredi, Ricardo B.
    Summers, Ronald M.
    RADIOLOGY, 2023, 309 (01)
  • [6] Assessing breast cancer chemotherapy response in radiology and pathology reports via a Large Language Model
    Dodhia, Parth
    Meepagala, Shawn
    Moallem, Golanz
    Rubin, Daniel
    Bean, Gregory
    Rusu, Mirabela
    IMAGING INFORMATICS FOR HEALTHCARE, RESEARCH, AND APPLICATIONS, MEDICAL IMAGING 2024, 2024, 12931
  • [7] Comparative Evaluation of Large Language Models for Translating Radiology Reports into Hindi
    Gupta, Amit
    Rastogi, Ashish
    Malhotra, Hema
    Rangarajan, Krithika
    INDIAN JOURNAL OF RADIOLOGY AND IMAGING, 2024,
  • [8] From jargon to clarity: Improving the readability of foot and ankle radiology reports with an artificial intelligence large language model
    Butler, James J.
    Harrington, Michael C.
    Tong, Yixuan
    Rosenbaum, Andrew J.
    Samsonov, Alan P.
    Walls, Raymond J.
    Kennedy, John G.
    FOOT AND ANKLE SURGERY, 2024, 30 (04) : 331 - 337
  • [9] The Fine-Tuned Large Language Model for Extracting the Progressive Bone Metastasis from Unstructured Radiology Reports
    Kanemaru, Noriko
    Yasaka, Koichiro
    Fujita, Nana
    Kanzawa, Jun
    Abe, Osamu
    JOURNAL OF IMAGING INFORMATICS IN MEDICINE, 2024, : 865 - 872
  • [10] Monitoring Patients with Glioblastoma by Using a Large Language Model: Accurate Summarization of Radiology Reports with GPT-4
    Laukamp, Kai R.
    Terzis, Robert A.
    Werner, Jan-Michael
    Galldiks, Norbert
    Lennartz, Simon
    Maintz, David
    Reimer, Robert
    Fervers, Philipp
    Gertz, Roman Johannes
    Persigehl, Thorsten
    Rubbert, Christian
    Lehnen, Nils C.
    Deuschl, Cornelius
    Schlamann, Marc
    Schoenfeld, Michael H.
    Kottlors, Jonathan
    RADIOLOGY, 2024, 312 (01)