Detecting insertion, substitution, and deletion errors in radiology reports using neural sequence-to-sequence models

被引:9
|
作者
Zech, John [1 ]
Forde, Jessica [2 ]
Titano, Joseph J. [1 ]
Kaji, Deepak [3 ]
Costa, Anthony [3 ]
Oermann, Eric Karl [3 ]
机构
[1] Icahn Sch Med, Dept Radiol, New York, NY USA
[2] 190 Doe Lib, Project Jupyter, Berkeley, CA USA
[3] Icahn Sch Med, Dept Neurosurg, New York, NY USA
关键词
Radiology; natural language processing; artificial intelligence; machine learning; neural networks (computer); SPEECH RECOGNITION; VOICE RECOGNITION; RATES;
D O I
10.21037/atm.2018.08.11
中图分类号
R73 [肿瘤学];
学科分类号
100214 ;
摘要
Background: Errors in grammar, spelling, and usage in radiology reports are common. To automatically detect inappropriate insertions, deletions, and substitutions of words in radiology reports, we proposed using a neural sequence-to-sequence (seq2seq) model. Methods: Head CT and chest radiograph reports from Mount Sinai Hospital (MSH) (n=61,722 and 818,978, respectively), Mount Sinai Queens (MSQ) (n=30,145 and 194,309, respectively) and MIMIC-III (n=32,259 and 54,685) were converted into sentences. Insertions, substitutions, and deletions of words were randomly introduced. Seq2seq models were trained using corrupted sentences as input to predict original uncorrupted sentences. Three models were trained using head CTs from MSH, chest radiographs from MSH, and head CTs from all three collections. Model performance was assessed across different sites and modalities. A sample of original, uncorrupted sentences were manually reviewed for any error in syntax, usage, or spelling to estimate real-world proofreading performance of the algorithm. Results: Seq2seq detected 90.3% and 88.2% of corrupted sentences with 97.7% and 98.8% specificity in same-site, same-modality test sets for head CTs and chest radiographs, respectively. Manual review of original, uncorrupted same-site same-modality head CT sentences demonstrated seq2seq positive predictive value (PPV) 0.393 (157/400; 95% CI, 0.346-0.441) and negative predictive value (NPV) 0.986 (789/800; 95% CI, 0.976-0.992) for detecting sentences containing real-world errors, with estimated sensitivity of 0.389 (95% CI, 0.267-0.542) and specificity 0.986 (95% CI, 0.985-0.987) over n=86,211 uncorrupted training examples. Conclusions: Seq2seq models can be highly effective at detecting erroneous insertions, deletions, and substitutions of words in radiology reports. To achieve high performance, these models require site- and modality-specific training examples. Incorporating additional targeted training data could further improve performance in detecting real-world errors in reports.
引用
收藏
页数:9
相关论文
共 50 条
  • [21] Learning Neural Sequence-to-Sequence Models from Weak Feedback with Bipolar Ramp Loss
    Jehl, Laura
    Lawrence, Carotin
    Riezler, Stefan
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2019, 7 : 233 - 248
  • [22] Identifying cause-and-effect relationships of manufacturing errors using sequence-to-sequence learning
    Reimer, Jeff
    Wang, Yandong
    Laridi, Sofiane
    Urdich, Juergen
    Wilmsmeier, Soeren
    Palmer, Gregory
    SCIENTIFIC REPORTS, 2022, 12 (01)
  • [23] Plasma confinement mode classification using a sequence-to-sequence neural network with attention
    Matos, F.
    Menkovski, V.
    Pau, A.
    Marceca, G.
    Jenko, F.
    NUCLEAR FUSION, 2021, 61 (04)
  • [24] Turkish Data-to-Text Generation Using Sequence-to-Sequence Neural Networks
    Demir, Seniz
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2023, 22 (02)
  • [25] Anomaly Detection for Industrial Control Systems Using Sequence-to-Sequence Neural Networks
    Kim, Jonguk
    Yun, Jeong-Han
    Kim, Hyoung Chun
    COMPUTER SECURITY, ESORICS 2019, 2020, 11980 : 3 - 18
  • [26] STREAMING SMALL-FOOTPRINT KEYWORD SPOTTING USING SEQUENCE-TO-SEQUENCE MODELS
    He, Yanzhang
    Prabhavalkar, Rohit
    Rao, Kanishka
    Li, Wei
    Bakhtin, Anton
    McGraw, Ian
    2017 IEEE AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING WORKSHOP (ASRU), 2017, : 474 - 481
  • [27] Integrated Training for Sequence-to-Sequence Models Using Non-Autoregressive Transformer
    Tokarchuk, Evgeniia
    Rosendahl, Jan
    Wang, Weiyue
    Petrushkov, Pavel
    Lancewicki, Tomer
    Khadivi, Shahram
    Ney, Hermann
    IWSLT 2021: THE 18TH INTERNATIONAL CONFERENCE ON SPOKEN LANGUAGE TRANSLATION, 2021, : 276 - 286
  • [28] Sequence-To-Sequence Neural Networks Inference on Embedded Processors Using Dynamic Beam Search
    Pagliari, Daniele Jahier
    Daghero, Francesco
    Poncino, Massimo
    ELECTRONICS, 2020, 9 (02)
  • [29] USING LOCAL PHRASE DEPENDENCY STRUCTURE INFORMATION IN NEURAL SEQUENCE-TO-SEQUENCE SPEECH SYNTHESIS
    Kaiki, Nobuyoshi
    Sakti, Sakriani
    Nakamura, Satoshi
    2021 24TH CONFERENCE OF THE ORIENTAL COCOSDA INTERNATIONAL COMMITTEE FOR THE CO-ORDINATION AND STANDARDISATION OF SPEECH DATABASES AND ASSESSMENT TECHNIQUES (O-COCOSDA), 2021, : 206 - 211
  • [30] Tagging Malware Intentions by Using Attention-Based Sequence-to-Sequence Neural Network
    Huang, Yi-Ting
    Chen, Yu-Yuan
    Yang, Chih-Chun
    Sun, Yeali
    Hsiao, Shun-Wen
    Chen, Meng Chang
    INFORMATION SECURITY AND PRIVACY, ACISP 2019, 2019, 11547 : 660 - 668