Towards Robust Neural Machine Reading Comprehension via Question Paraphrases

被引:0
|
作者
Li, Ying [1 ]
Li, Hongyu [2 ]
Liu, Jing [2 ]
机构
[1] Univ Sci & Technol China, Natl Engn Lab Brain Inspired Intelligence Technol, Hefei, Peoples R China
[2] Baidu Inc, Beijing, Peoples R China
关键词
machine reading comprehension; oversensitivity; question paraphrases;
D O I
10.1109/ialp48816.2019.9037673
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we focus on addressing the oversensitivity issue of neural machine reading comprehension (MRC) models. By oversensitivity, we mean that the neural MRC models give different answers to question paraphrases that are semantically equivalent. To address this issue, we first create a large-scale Chinese MRC dataset with high quality question paraphrases generated by a toolkit used in Baidu Search. Then, we quantitively analyze the oversensitivity issue of the neural MRC models on the dataset. Intuitively, if two questions are paraphrases of each other, a robust model should give the same predictions. Based on this intuition, we propose a regularized BERT-based model to encourage the model give the same predictions to similar inputs by lever-aging high-quality question paraphrases. The experimental results show that our approaches can significantly improve the robustness of a strong BERT-based MRC model and achieve improvements over the BERT-based model in terms of held-out accuracy. Specifically, the different prediction ratio (DPR) for question paraphrases of the proposed model decreases more than 10%.
引用
收藏
页码:290 / 295
页数:6
相关论文
共 50 条
  • [41] A Multi-Stage Memory Augmented Neural Network for Machine Reading Comprehension
    Yu, Seunghak
    Indurthi, Sathish
    Back, Seohyun
    Lee, Haejun
    MACHINE READING FOR QUESTION ANSWERING, 2018, : 21 - 30
  • [42] Extraction of Question-related Sentences for Reading Comprehension Tests via Attention Mechanism
    Shan, Junjie
    Nishihara, Yoko
    Maeda, Akira
    Yamanishi, Ryosuke
    2020 25TH INTERNATIONAL CONFERENCE ON TECHNOLOGIES AND APPLICATIONS OF ARTIFICIAL INTELLIGENCE (TAAI 2020), 2020, : 23 - 28
  • [43] Understand before Answer: Improve Temporal Reading Comprehension via Precise Question Understanding
    Huang, Hao
    Geng, Xiubo
    Long, Guodong
    Jiang, Daxin
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 385 - 394
  • [44] Zero-Shot Slot Filling via Latent Question Representation and Reading Comprehension
    Wu, Tongtong
    Wang, Meng
    Gao, Huan
    Qi, Guilin
    Li, Weizhuo
    PRICAI 2019: TRENDS IN ARTIFICIAL INTELLIGENCE, PT III, 2019, 11672 : 123 - 136
  • [45] NER-MQMRC: Formulating Named Entity Recognition as Multi Question Machine Reading Comprehension
    Shrimal, Anubhav
    Jain, Avi
    Mehta, Kartik
    Yenigalla, Promod
    2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, NAACL-HLT 2022, 2022, : 230 - 238
  • [46] A Survey on Machine Reading Comprehension Systems
    Baradaran, Razieh
    Ghiasi, Razieh
    Amirkhani, Hossein
    NATURAL LANGUAGE ENGINEERING, 2022, 28 (06) : 683 - 732
  • [47] Event Extraction as Machine Reading Comprehension
    Liu, Jian
    Chen, Yubo
    Liu, Kang
    Bi, Wei
    Liu, Xiaojiang
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 1641 - 1651
  • [48] Improving Machine Reading Comprehension with General Reading Strategies
    Sun, Kai
    Yu, Dian
    Yu, Dong
    Cardie, Claire
    2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, 2019, : 2633 - 2643
  • [49] Machine Reading Comprehension: Matching and Orders
    Liu, Ao
    Qu, Lizhen
    Lu, Junyu
    Zhang, Chenbin
    Xu, Zenglin
    PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT (CIKM '19), 2019, : 2057 - 2060
  • [50] Reading Comprehension in Czech via Machine Translation and Cross-Lingual Transfer
    Mackova, Katerina
    Straka, Milan
    TEXT, SPEECH, AND DIALOGUE (TSD 2020), 2020, 12284 : 171 - 179