Detecting insertion, substitution, and deletion errors in radiology reports using neural sequence-to-sequence models

被引:9
|
作者
Zech, John [1 ]
Forde, Jessica [2 ]
Titano, Joseph J. [1 ]
Kaji, Deepak [3 ]
Costa, Anthony [3 ]
Oermann, Eric Karl [3 ]
机构
[1] Icahn Sch Med, Dept Radiol, New York, NY USA
[2] 190 Doe Lib, Project Jupyter, Berkeley, CA USA
[3] Icahn Sch Med, Dept Neurosurg, New York, NY USA
关键词
Radiology; natural language processing; artificial intelligence; machine learning; neural networks (computer); SPEECH RECOGNITION; VOICE RECOGNITION; RATES;
D O I
10.21037/atm.2018.08.11
中图分类号
R73 [肿瘤学];
学科分类号
100214 ;
摘要
Background: Errors in grammar, spelling, and usage in radiology reports are common. To automatically detect inappropriate insertions, deletions, and substitutions of words in radiology reports, we proposed using a neural sequence-to-sequence (seq2seq) model. Methods: Head CT and chest radiograph reports from Mount Sinai Hospital (MSH) (n=61,722 and 818,978, respectively), Mount Sinai Queens (MSQ) (n=30,145 and 194,309, respectively) and MIMIC-III (n=32,259 and 54,685) were converted into sentences. Insertions, substitutions, and deletions of words were randomly introduced. Seq2seq models were trained using corrupted sentences as input to predict original uncorrupted sentences. Three models were trained using head CTs from MSH, chest radiographs from MSH, and head CTs from all three collections. Model performance was assessed across different sites and modalities. A sample of original, uncorrupted sentences were manually reviewed for any error in syntax, usage, or spelling to estimate real-world proofreading performance of the algorithm. Results: Seq2seq detected 90.3% and 88.2% of corrupted sentences with 97.7% and 98.8% specificity in same-site, same-modality test sets for head CTs and chest radiographs, respectively. Manual review of original, uncorrupted same-site same-modality head CT sentences demonstrated seq2seq positive predictive value (PPV) 0.393 (157/400; 95% CI, 0.346-0.441) and negative predictive value (NPV) 0.986 (789/800; 95% CI, 0.976-0.992) for detecting sentences containing real-world errors, with estimated sensitivity of 0.389 (95% CI, 0.267-0.542) and specificity 0.986 (95% CI, 0.985-0.987) over n=86,211 uncorrupted training examples. Conclusions: Seq2seq models can be highly effective at detecting erroneous insertions, deletions, and substitutions of words in radiology reports. To achieve high performance, these models require site- and modality-specific training examples. Incorporating additional targeted training data could further improve performance in detecting real-world errors in reports.
引用
收藏
页数:9
相关论文
共 50 条
  • [41] Stochastic models of sequence evolution including insertion-deletion events
    Miklos, Istvan
    Novak, Adam
    Satija, Rahul
    Lyngso, Rune
    Hein, Jotun
    STATISTICAL METHODS IN MEDICAL RESEARCH, 2009, 18 (05) : 453 - 485
  • [42] EXTRACTING UNIT EMBEDDINGS USING SEQUENCE-TO-SEQUENCE ACOUSTIC MODELS FOR UNIT SELECTION SPEECH SYNTHESIS
    Zhou, Xiao
    Ling, Zhen-Hua
    Dai, Li-Rong
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 7659 - 7663
  • [43] Extracting templates from radiology reports using sequence alignment
    Wu, Shengyang
    Langlotz, Curtis P.
    Lakhani, Paras
    Ungar, Lyle H.
    INTERNATIONAL JOURNAL OF DATA MINING AND BIOINFORMATICS, 2012, 6 (06) : 633 - 650
  • [44] Extracting Templates from Radiology Reports using Sequence Alignment
    Wu, Shengyang
    Langlotz, Curtis P.
    Lakhani, Paras
    Ungar, Lyle H.
    BIBMW: 2009 IEEE INTERNATIONAL CONFERENCE ON BIOINFORMATICS AND BIOMEDICINE WORKSHOP, 2009, : 314 - 318
  • [45] Classification of Radiology Reports Using Neural Attention Models
    Shin, Bonggun
    Chokshi, Falgun H.
    Lee, Timothy
    Choi, Jinho D.
    2017 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2017, : 4363 - 4370
  • [46] Fault diagnostics on railway switches and crossings using a bat-optimised sequence-to-sequence neural model
    Deng, Xiaoyun
    INTERNATIONAL JOURNAL OF SENSOR NETWORKS, 2024, 44 (04) : 226 - 236
  • [47] Data2Vis: Automatic Generation of Data Visualizations Using Sequence-to-Sequence Recurrent Neural Networks
    Dibia, Victor
    Demiralp, Cagatay
    IEEE COMPUTER GRAPHICS AND APPLICATIONS, 2019, 39 (05) : 33 - 46
  • [48] Non-verbal Vocalisation and Laughter Detection using Sequence-to-sequence Models and Multi-label Training
    Condron, Scott
    Clarke, Georgia
    Klementiev, Anita
    Morse-Kopp, Daniela
    Parry, Jack
    Palaz, Dimitri
    INTERSPEECH 2021, 2021, : 2506 - 2510
  • [49] Abstractive Text Summarization: Enhancing Sequence-to-Sequence Models Using Word Sense Disambiguation and Semantic Content Generalization
    Kouris, Panagiotis
    Alexandridis, Georgios
    Stafylopatis, Andreas
    COMPUTATIONAL LINGUISTICS, 2021, 47 (04) : 813 - 859
  • [50] Sequence-to-Sequence Learning-Based Conversion of Pseudo-Code to Source Code Using Neural Translation Approach
    Acharjee, Uzzal Kumar
    Arefin, Minhazul
    Hossen, Kazi Mojammel
    Uddin, Mohammed Nasir
    Uddin, Md Ashraf
    Islam, Linta
    IEEE ACCESS, 2022, 10 : 26730 - 26742