Evaluating Neural Model Robustness for Machine Comprehension

被引:0
|
作者
Wu, Winston [1 ]
Arendt, Dustin [2 ]
Volkova, Svitlana [3 ]
机构
[1] Johns Hopkins Univ, Dept Comp Sci, Ctr Language & Speech Proc, Baltimore, MD 21218 USA
[2] Pacific Northwest Natl Lab, Visual Analyt Grp, Richland, WA USA
[3] Pacific Northwest Natl Lab, Data Sci & Analyt Grp, Richland, WA USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We evaluate neural model robustness to adversarial attacks using different types of linguistic unit perturbations - character and word, and propose a new method for strategic sentencelevel perturbations. We experiment with different amounts of perturbations to examine model confidence and misclassification rate, and contrast model performance with different embeddings BERT and ELMo on two benchmark datasets SQuAD and TriviaQA. We demonstrate how to improve model performance during an adversarial attack by using ensembles. Finally, we analyze factors that affect model behavior under adversarial attack, and develop a new model to predict errors during attacks. Our novel findings reveal that (a) unlike BERT, models that use ELMo embeddings are more susceptible to adversarial attacks, (b) unlike word and paraphrase, character perturbations affect the model the most but are most easily compensated for by adversarial training, (c) word perturbations lead to more high-confidence misclassifications compared to sentence- and character-level perturbations, (d) the type of question and model answer length (the longer the answer the more likely it is to be incorrect) is the most predictive of model errors in adversarial setting, and (e) conclusions about model behavior are dataset-specific.
引用
收藏
页码:2470 / 2481
页数:12
相关论文
共 50 条
  • [21] Evaluating the Robustness of Neural Language Models to Input Perturbations
    Moradi, Milad
    Samwald, Matthias
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 1558 - 1570
  • [22] Evaluating the Robustness of Ultrasound Beamforming with Deep Neural Networks
    Luchies, Adam
    Byram, Brett
    2018 IEEE INTERNATIONAL ULTRASONICS SYMPOSIUM (IUS), 2018,
  • [23] Survey of Machine Reading Comprehension Based on Neural Network
    Gu Y.-J.
    Gui X.-L.
    Li D.-F.
    Shen Y.
    Liao D.
    Ruan Jian Xue Bao/Journal of Software, 2020, 31 (07): : 2095 - 2126
  • [24] XCMRC: Evaluating Cross-Lingual Machine Reading Comprehension
    Liu, Pengyuan
    Deng, Yuning
    Zhu, Chenghao
    Hu, Han
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING (NLPCC 2019), PT I, 2019, 11838 : 552 - 564
  • [25] SENTENCE BOUNDARY AUGMENTATION FOR NEURAL MACHINE TRANSLATION ROBUSTNESS
    Li, Daniel
    Te, I
    Arivazhagan, Naveen
    Cherry, Colin
    Padfield, Dirk
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 7553 - 7557
  • [26] VS3-NET: Neural variational inference model for machine-reading comprehension
    Park, Cheoneum
    Lee, Changki
    Song, Heelmn
    ETRI JOURNAL, 2019, 41 (06) : 771 - 781
  • [27] Robustness of a neural network model for differencing
    Solodovnikov, A
    Reed, MC
    JOURNAL OF COMPUTATIONAL NEUROSCIENCE, 2001, 11 (02) : 165 - 173
  • [28] Evaluating Model Robustness and Stability to Dataset Shift
    Subbaswamy, Adarsh
    Adams, Roy
    Saria, Suchi
    24TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS (AISTATS), 2021, 130
  • [29] Robustness of a Neural Network Model for Differencing
    Alexander Solodovnikov
    Michael C. Reed
    Journal of Computational Neuroscience, 2001, 11 : 165 - 173
  • [30] A Hierarchical Iterative Attention Model for Machine Comprehension
    Liu, Zhuang
    Huang, Degen
    Zhang, Yunxia
    Zhang, Chuan
    NATURAL LANGUAGE PROCESSING AND INFORMATION SYSTEMS, NLDB 2017, 2017, 10260 : 341 - 352