Enhancing Pre-Trained Language Representations with Rich Knowledge for Machine Reading Comprehension

被引:0
|
作者
Yang, An [1 ,2 ]
Wang, Quan [2 ]
Liu, Jing [2 ]
Liu, Kai [2 ]
Lyu, Yajuan [2 ]
Wu, Hua [2 ]
She, Qiaoqiao [2 ]
Li, Sujian [1 ]
机构
[1] Peking Univ, MOE, Key Lab Computat Linguist, Beijing, Peoples R China
[2] Baidu Inc, Beijing, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Machine reading comprehension (MRC) is a crucial and challenging task in NLP. Recently, pre-trained language models (LMs), especially BERT, have achieved remarkable success, presenting new state-of-the-art results in MRC. In this work, we investigate the potential of leveraging external knowledge bases (KBs) to further improve BERT for MRC. We introduce KT-NET, which employs an attention mechanism to adaptively select desired knowledge from KBs, and then fuses selected knowledge with BERT to enable context- and knowledge-aware predictions. We believe this would combine the merits of both deep LMs and curated KBs towards better MRC. Experimental results indicate that KT-NET offers significant and consistent improvements over BERT, outperforming competitive baselines on ReCoRD and SQuAD1.1 benchmarks. Notably, it ranks the 1st place on the ReCoRD leaderboard, and is also the best single model on the SQuAD1.1 leaderboard at the time of submission (March 4th, 2019).(1)
引用
收藏
页码:2346 / 2357
页数:12
相关论文
共 50 条
  • [1] Research on machine reading comprehension based on pre-trained model
    Chen, Guanlin
    Yao, Rutao
    Zhou, Haiwei
    Li, Tian
    Yang, Wujian
    [J]. International Journal of Reasoning-based Intelligent Systems, 2022, 14 (04): : 240 - 246
  • [2] From Cloze to Comprehension: Retrofitting Pre-trained Masked Language Models to Pre-trained Machine Reader
    Xu, Weiwen
    Li, Xin
    Zhang, Wenxuan
    Zhou, Meng
    Lam, Wai
    Si, Luo
    Bing, Lidong
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [3] Exploiting Diverse Information in Pre-Trained Language Model for Multi-Choice Machine Reading Comprehension
    Bai, Ziwei
    Liu, Junpeng
    Wang, Meiqi
    Yuan, Caixia
    Wang, Xiaojie
    [J]. APPLIED SCIENCES-BASEL, 2022, 12 (06):
  • [4] HORNET: Enriching Pre-trained Language Representations with Heterogeneous Knowledge Sources
    Zhang, Taolin
    Cai, Zerui
    Wang, Chengyu
    Li, Peng
    Li, Yang
    Qiu, Minghui
    Tang, Chengguang
    He, Xiaofeng
    Huang, Jun
    [J]. PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 2608 - 2617
  • [5] Pre-trained Language Model Representations for Language Generation
    Edunov, Sergey
    Baevski, Alexei
    Auli, Michael
    [J]. 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, 2019, : 4052 - 4059
  • [6] Graph-combined Coreference Resolution Methods on Conversational Machine Reading Comprehension with Pre-trained Language Model
    Wang, Zhaodong
    Komatani, Kazunori
    [J]. PROCEEDINGS OF THE SECOND DIALDOC WORKSHOP ON DOCUMENT-GROUNDED DIALOGUE AND CONVERSATIONAL QUESTION ANSWERING (DIALDOC 2022), 2022, : 72 - 82
  • [7] On the Language Neutrality of Pre-trained Multilingual Representations
    Libovicky, Jindrich
    Rosa, Rudolf
    Fraser, Alexander
    [J]. FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 1663 - 1674
  • [8] Evaluating the Summarization Comprehension of Pre-Trained Language Models
    Chernyshev, D. I.
    Dobrov, B. V.
    [J]. LOBACHEVSKII JOURNAL OF MATHEMATICS, 2023, 44 (08) : 3028 - 3039
  • [9] Evaluating the Summarization Comprehension of Pre-Trained Language Models
    D. I. Chernyshev
    B. V. Dobrov
    [J]. Lobachevskii Journal of Mathematics, 2023, 44 : 3028 - 3039
  • [10] Machine Reading Comprehension with Rich Knowledge
    He, Jun
    Peng, Li
    Zhang, Yinghui
    Sun, Bo
    Xiao, Rong
    Xiao, Yongkang
    [J]. INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2022, 36 (05)