Detecting insertion, substitution, and deletion errors in radiology reports using neural sequence-to-sequence models

被引:9
|
作者
Zech, John [1 ]
Forde, Jessica [2 ]
Titano, Joseph J. [1 ]
Kaji, Deepak [3 ]
Costa, Anthony [3 ]
Oermann, Eric Karl [3 ]
机构
[1] Icahn Sch Med, Dept Radiol, New York, NY USA
[2] 190 Doe Lib, Project Jupyter, Berkeley, CA USA
[3] Icahn Sch Med, Dept Neurosurg, New York, NY USA
关键词
Radiology; natural language processing; artificial intelligence; machine learning; neural networks (computer); SPEECH RECOGNITION; VOICE RECOGNITION; RATES;
D O I
10.21037/atm.2018.08.11
中图分类号
R73 [肿瘤学];
学科分类号
100214 ;
摘要
Background: Errors in grammar, spelling, and usage in radiology reports are common. To automatically detect inappropriate insertions, deletions, and substitutions of words in radiology reports, we proposed using a neural sequence-to-sequence (seq2seq) model. Methods: Head CT and chest radiograph reports from Mount Sinai Hospital (MSH) (n=61,722 and 818,978, respectively), Mount Sinai Queens (MSQ) (n=30,145 and 194,309, respectively) and MIMIC-III (n=32,259 and 54,685) were converted into sentences. Insertions, substitutions, and deletions of words were randomly introduced. Seq2seq models were trained using corrupted sentences as input to predict original uncorrupted sentences. Three models were trained using head CTs from MSH, chest radiographs from MSH, and head CTs from all three collections. Model performance was assessed across different sites and modalities. A sample of original, uncorrupted sentences were manually reviewed for any error in syntax, usage, or spelling to estimate real-world proofreading performance of the algorithm. Results: Seq2seq detected 90.3% and 88.2% of corrupted sentences with 97.7% and 98.8% specificity in same-site, same-modality test sets for head CTs and chest radiographs, respectively. Manual review of original, uncorrupted same-site same-modality head CT sentences demonstrated seq2seq positive predictive value (PPV) 0.393 (157/400; 95% CI, 0.346-0.441) and negative predictive value (NPV) 0.986 (789/800; 95% CI, 0.976-0.992) for detecting sentences containing real-world errors, with estimated sensitivity of 0.389 (95% CI, 0.267-0.542) and specificity 0.986 (95% CI, 0.985-0.987) over n=86,211 uncorrupted training examples. Conclusions: Seq2seq models can be highly effective at detecting erroneous insertions, deletions, and substitutions of words in radiology reports. To achieve high performance, these models require site- and modality-specific training examples. Incorporating additional targeted training data could further improve performance in detecting real-world errors in reports.
引用
收藏
页数:9
相关论文
共 50 条
  • [1] Retrosynthetic Reaction Prediction Using Neural Sequence-to-Sequence Models
    Liu, Bowen
    Ramsundar, Bharath
    Kawthekar, Prasad
    Shi, Jade
    Gomes, Joseph
    Quang Luu Nguyen
    Ho, Stephen
    Sloane, Jack
    Wender, Paul
    Pande, Vijay
    ACS CENTRAL SCIENCE, 2017, 3 (10) : 1103 - 1113
  • [2] Neural AMR: Sequence-to-Sequence Models for Parsing and Generation
    Konstas, Ioannis
    Iyer, Srinivasan
    Yatskar, Mark
    Choi, Yejin
    Zettlemoyer, Luke
    PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1, 2017, : 146 - 157
  • [3] Neural Abstractive Text Summarization with Sequence-to-Sequence Models
    Shi, Tian
    Keneshloo, Yaser
    Ramakrishnan, Naren
    Reddy, Chandan K.
    ACM/IMS Transactions on Data Science, 2021, 2 (01):
  • [4] Incremental Text to Speech for Neural Sequence-to-Sequence Models using Reinforcement Learning
    Mohan, Devang S. Ram
    Lenain, Raphael
    Foglianti, Lorenzo
    Teh, Tian Huey
    Staib, Marlene
    Torresquintero, Alexandra
    Gao, Jiameng
    INTERSPEECH 2020, 2020, : 3186 - 3190
  • [5] Generating Medical Reports from Patient-Doctor Conversations using Sequence-to-Sequence Models
    Enarvi, Seppo
    Amoia, Marilisa
    Teba, Miguel Del-Agua
    Delaney, Brian
    Diehl, Frank
    Gallopyn, Guido
    Hahn, Stefan
    Harris, Kristina
    McGrath, Liam
    Pan, Yue
    Pinto, Joel
    Rubini, Luca
    Ruiz, Miguel
    Singh, Gagandeep
    Stemmer, Fabian
    Sun, Weiyi
    Vozila, Paul
    Lin, Thomas
    Ramamurthy, Ranjani
    NATURAL LANGUAGE PROCESSING FOR MEDICAL CONVERSATIONS, 2020, : 22 - 30
  • [6] Persian Keyphrase Generation Using Sequence-to-sequence Models
    Doostmohammadi, Ehsan
    Bokaei, Mohammad Hadi
    Sameti, Hossein
    2019 27TH IRANIAN CONFERENCE ON ELECTRICAL ENGINEERING (ICEE 2019), 2019, : 2010 - 2015
  • [7] Detecting multi-stage attacks using sequence-to-sequence model
    Zhou, Peng
    Zhou, Gongyan
    Wu, Dakui
    Fei, Minrui
    COMPUTERS & SECURITY, 2021, 105
  • [8] Sequence-to-Sequence Neural Net Models for Grapheme-to-Phoneme Conversion
    Yao, Kaisheng
    Zweig, Geoffrey
    16TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2015), VOLS 1-5, 2015, : 3330 - 3334
  • [9] Reformulating natural language queries using sequence-to-sequence models
    Xiaoyu Liu
    Shunda Pan
    Qi Zhang
    Yu-Gang Jiang
    Xuanjing Huang
    Science China Information Sciences, 2019, 62
  • [10] Runoff predictions in ungauged basins using sequence-to-sequence models
    Yin, Hanlin
    Guo, Zilong
    Zhang, Xiuwei
    Chen, Jiaojiao
    Zhang, Yanning
    JOURNAL OF HYDROLOGY, 2021, 603