Towards Robust Neural Machine Reading Comprehension via Question Paraphrases

被引:0
|
作者
Li, Ying [1 ]
Li, Hongyu [2 ]
Liu, Jing [2 ]
机构
[1] Univ Sci & Technol China, Natl Engn Lab Brain Inspired Intelligence Technol, Hefei, Peoples R China
[2] Baidu Inc, Beijing, Peoples R China
关键词
machine reading comprehension; oversensitivity; question paraphrases;
D O I
10.1109/ialp48816.2019.9037673
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we focus on addressing the oversensitivity issue of neural machine reading comprehension (MRC) models. By oversensitivity, we mean that the neural MRC models give different answers to question paraphrases that are semantically equivalent. To address this issue, we first create a large-scale Chinese MRC dataset with high quality question paraphrases generated by a toolkit used in Baidu Search. Then, we quantitively analyze the oversensitivity issue of the neural MRC models on the dataset. Intuitively, if two questions are paraphrases of each other, a robust model should give the same predictions. Based on this intuition, we propose a regularized BERT-based model to encourage the model give the same predictions to similar inputs by lever-aging high-quality question paraphrases. The experimental results show that our approaches can significantly improve the robustness of a strong BERT-based MRC model and achieve improvements over the BERT-based model in terms of held-out accuracy. Specifically, the different prediction ratio (DPR) for question paraphrases of the proposed model decreases more than 10%.
引用
收藏
页码:290 / 295
页数:6
相关论文
共 50 条
  • [21] Frame-based Neural Network for Machine Reading Comprehension
    Guo, Shaoru
    Guan, Yong
    Tan, Hongye
    Li, Ru
    Li, Xiaoli
    KNOWLEDGE-BASED SYSTEMS, 2021, 219 (219)
  • [22] Design of A Recurrent Neural Network Model for Machine Reading Comprehension
    Singh, Uttam
    Kedas, Shweta
    Prasanth, Sikakollu
    Kumar, Arun
    Semwal, Vijay Bhaskar
    Tikkiwal, Vinay Anand
    INTERNATIONAL CONFERENCE ON COMPUTATIONAL INTELLIGENCE AND DATA SCIENCE, 2020, 167 : 1791 - 1800
  • [23] Robust Reading Comprehension With Linguistic Constraints via Posterior Regularization
    Zhou, Mantong
    Huang, Minlie
    Zhu, Xiaoyan
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2020, 28 : 2500 - 2510
  • [24] Improving the robustness of machine reading comprehension via contrastive learning
    Jianzhou Feng
    Jiawei Sun
    Di Shao
    Jinman Cui
    Applied Intelligence, 2023, 53 : 9103 - 9114
  • [25] Improving the robustness of machine reading comprehension via contrastive learning
    Feng, Jianzhou
    Sun, Jiawei
    Shao, Di
    Cui, Jinman
    APPLIED INTELLIGENCE, 2023, 53 (08) : 9103 - 9114
  • [26] DAQAS: Deep Arabic Question Answering System based on duplicate question detection and machine reading comprehension
    Alami, Hamza
    Mahdaouy, Abdelkader El
    Benlahbib, Abdessamad
    En-Nahnahi, Noureddine
    Berrada, Ismail
    Ouatik, Said El Alaoui
    JOURNAL OF KING SAUD UNIVERSITY-COMPUTER AND INFORMATION SCIENCES, 2023, 35 (08)
  • [27] Towards Medical Machine Reading Comprehension with Structural Knowledge and Plain Text
    Li, Dongfang
    Hu, Baotian
    Chen, Qingcai
    Peng, Weihua
    Wang, Anqi
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 1427 - 1438
  • [28] Visual Question Answering as Reading Comprehension
    Li, Hui
    Wang, Peng
    Shen, Chunhua
    van den Hengel, Anton
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 6312 - 6321
  • [29] Chinese machine reading comprehension based on deep learning neural network
    Ma, Chao
    An, Jing
    Xu, Jing
    Xu, Binchen
    Xu, Luyuan
    Bai, Xiang-En
    INTERNATIONAL JOURNAL OF BIO-INSPIRED COMPUTATION, 2023, 21 (03) : 137 - 147
  • [30] Survey on Machine Reading Comprehension
    Wang X.-J.
    Bai Z.-W.
    Li K.
    Yuan C.-X.
    Beijing Youdian Daxue Xuebao/Journal of Beijing University of Posts and Telecommunications, 2019, 42 (06): : 1 - 9