Enhancing Pre-Trained Language Representations with Rich Knowledge for Machine Reading Comprehension

被引:0
|
作者
Yang, An [1 ,2 ]
Wang, Quan [2 ]
Liu, Jing [2 ]
Liu, Kai [2 ]
Lyu, Yajuan [2 ]
Wu, Hua [2 ]
She, Qiaoqiao [2 ]
Li, Sujian [1 ]
机构
[1] Peking Univ, MOE, Key Lab Computat Linguist, Beijing, Peoples R China
[2] Baidu Inc, Beijing, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Machine reading comprehension (MRC) is a crucial and challenging task in NLP. Recently, pre-trained language models (LMs), especially BERT, have achieved remarkable success, presenting new state-of-the-art results in MRC. In this work, we investigate the potential of leveraging external knowledge bases (KBs) to further improve BERT for MRC. We introduce KT-NET, which employs an attention mechanism to adaptively select desired knowledge from KBs, and then fuses selected knowledge with BERT to enable context- and knowledge-aware predictions. We believe this would combine the merits of both deep LMs and curated KBs towards better MRC. Experimental results indicate that KT-NET offers significant and consistent improvements over BERT, outperforming competitive baselines on ReCoRD and SQuAD1.1 benchmarks. Notably, it ranks the 1st place on the ReCoRD leaderboard, and is also the best single model on the SQuAD1.1 leaderboard at the time of submission (March 4th, 2019).(1)
引用
收藏
页码:2346 / 2357
页数:12
相关论文
共 50 条
  • [41] Enhancing Chinese Pre-trained Language Model via Heterogeneous Linguistics Graph
    Li, Yanzeng
    Cao, Jiangxia
    Cong, Xin
    Zhang, Zhenyu
    Yu, Bowen
    Zhu, Hongsong
    Liu, Tingwen
    [J]. PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 1986 - 1996
  • [42] Enhancing Transformer-based language models with commonsense representations for knowledge-driven machine comprehension
    Li, Ronghan
    Jiang, Zejun
    Wang, Lifang
    Lu, Xinyu
    Zhao, Meng
    Chen, Daqing
    [J]. KNOWLEDGE-BASED SYSTEMS, 2021, 220
  • [43] Towards Knowledge Enhanced Language Model for Machine Reading Comprehension
    Gong, Peizhu
    Liu, Jin
    Yang, Yihe
    He, Huihua
    [J]. IEEE ACCESS, 2020, 8 : 224837 - 224851
  • [44] Improving Quality Estimation of Machine Translation by Using Pre-trained Language Representation
    Miao, Guoyi
    Di, Hui
    Xu, Jinan
    Yang, Zhongcheng
    Chen, Yufeng
    Ouchi, Kazushige
    [J]. MACHINE TRANSLATION, CCMT 2019, 2019, 1104 : 11 - 22
  • [45] ReAugKD: Retrieval-Augmented Knowledge Distillation For Pre-trained Language Models
    Zhang, Jianyi
    Muhamed, Aashiq
    Anantharaman, Aditya
    Wang, Guoyin
    Chen, Changyou
    Zhong, Kai
    Cui, Qingjun
    Xu, Yi
    Zeng, Belinda
    Chilimbi, Trishul
    Chen, Yiran
    [J]. 61ST CONFERENCE OF THE THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 2, 2023, : 1128 - 1136
  • [46] TwitterBERT: Framework for Twitter Sentiment Analysis Based on Pre-trained Language Model Representations
    Azzouza, Noureddine
    Akli-Astouati, Karima
    Ibrahim, Roliana
    [J]. EMERGING TRENDS IN INTELLIGENT COMPUTING AND INFORMATICS: DATA SCIENCE, INTELLIGENT INFORMATION SYSTEMS AND SMART COMPUTING, 2020, 1073 : 428 - 437
  • [47] Probing Multi-modal Machine Translation with Pre-trained Language Model
    Kong, Yawei
    Fan, Kai
    [J]. FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 3689 - 3699
  • [48] SimKGC: Simple Contrastive Knowledge Graph Completion with Pre-trained Language Models
    Wang, Liang
    Zhao, Wei
    Wei, Zhuoyu
    Liu, Jingming
    [J]. PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 4281 - 4294
  • [49] Assisted Process Knowledge Graph Building Using Pre-trained Language Models
    Bellan, Patrizio
    Dragoni, Mauro
    Ghidini, Chiara
    [J]. AIXIA 2022 - ADVANCES IN ARTIFICIAL INTELLIGENCE, 2023, 13796 : 60 - 74
  • [50] Exploring Accurate and Generic Simile Knowledge from Pre-trained Language Models
    Zhou, Shuhan
    Ma, Longxuan
    Shao, Yanqiu
    [J]. CHINESE COMPUTATIONAL LINGUISTICS, CCL 2023, 2023, 14232 : 348 - 363