Learning Knowledge-Enhanced Contextual Language Representations for Domain Natural Language Understanding

被引:0
|
作者
Zhang, Taolin [1 ,2 ]
Xu, Ruyao [1 ]
Wang, Chengyu [2 ]
Duan, Zhongjie [1 ]
Chen, Cen [1 ]
Qiu, Minghui [2 ]
Cheng, Dawei [3 ]
He, Xiaofeng [1 ]
Qian, Weining [1 ]
机构
[1] East China Normal Univ, Shanghai, Peoples R China
[2] Alibaba Grp, Hangzhou, Peoples R China
[3] Tongji Univ, Shanghai, Peoples R China
基金
中国国家自然科学基金;
关键词
MODEL;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Knowledge-Enhanced Pre-trained Language Models (KEPLMs) improve the performance of various downstream NLP tasks by injecting knowledge facts from large-scale Knowledge Graphs (KGs). However, existing methods for pre-training KEPLMs with relational triples are difficult to be adapted to close domains due to the lack of sufficient domain graph semantics. In this paper, we propose a Knowledgeenhanced lANGuAge Representation learning framework for various clOsed dOmains (KAN-GAROO) via capturing the implicit graph structure among the entities. Specifically, since the entity coverage rates of closed-domain KGs can be relatively low and may exhibit the global sparsity phenomenon for knowledge injection, we consider not only the shallow relational representations of triples but also the hyperbolic embeddings of deep hierarchical entityclass structures for effective knowledge fusion. Moreover, as two closed-domain entities under the same entity-class often have locally dense neighbor subgraphs counted by max point bi-connected component, we further propose a data augmentation strategy based on contrastive learning over subgraphs to construct hard negative samples of higher quality. It makes the underlying KELPMs better distinguish the semantics of these neighboring entities to further complement the global semantic sparsity. In the experiments, we evaluate KANGAROO over various knowledge-aware and general NLP tasks in both full and few-shot learning settings, outperforming various KEPLM training paradigms performance in closed-domains significantly.
引用
收藏
页码:15663 / 15676
页数:14
相关论文
共 50 条
  • [21] TexSmart: A System for Enhanced Natural Language Understanding
    Liu, Lemao
    Zhang, Haisong
    Jiang, Haiyun
    Li, Yangming
    Zhao, Enbo
    Xu, Kun
    Song, Linfeng
    Zheng, Suncong
    Zhou, Botong
    Zhu, Jianchen
    Feng, Xiao
    Chen, Tao
    Yang, Tao
    Yu, Dong
    Zhang, Feng
    Kang, Zhanhui
    Shi, Shuming
    ACL-IJCNLP 2021: THE JOINT CONFERENCE OF THE 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING: PROCEEDINGS OF THE SYSTEM DEMONSTRATIONS, 2021, : 1 - 10
  • [22] Knowledge-enhanced Prompt Learning for Open-domain Commonsense Reasoning
    Zhao, Xujiang
    Liu, Yanchi
    Cheng, Wei
    Oishi, Mika
    Osaki, Takao
    Matsuda, Katsushi
    Chen, Haifeng
    NEC Technical Journal, 2024, 17 (02): : 91 - 95
  • [23] Learning Unsupervised Knowledge-Enhanced Representations to Reduce the Semantic Gap in Information Retrieval
    Agosti, Maristella
    Marchesin, Stefano
    Silvello, Gianmaria
    ACM TRANSACTIONS ON INFORMATION SYSTEMS, 2020, 38 (04)
  • [24] A Software Bug Fixing Approach Based on Knowledge-Enhanced Large Language Models
    Bo, Lili
    He, Yuting
    Sun, Xiaobing
    Ji, Wangjie
    Wu, Xiaohan
    2024 IEEE 24TH INTERNATIONAL CONFERENCE ON SOFTWARE QUALITY, RELIABILITY AND SECURITY, QRS, 2024, : 169 - 179
  • [25] Design considerations for knowledge source representations of a stochastically-based natural language understanding component
    Minker, W
    SPEECH COMMUNICATION, 1999, 28 (02) : 141 - 154
  • [26] Robust Spoken Language Understanding with Acoustic and Domain Knowledge
    Li, Hao
    Liu, Chen
    Zhu, Su
    Yu, Kai
    ICMI'19: PROCEEDINGS OF THE 2019 INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2019, : 531 - 535
  • [27] Natural language processing and knowledge representation: Language for knowledge and knowledge for language
    Mercer, RE
    COMPUTATIONAL LINGUISTICS, 2001, 27 (02) : 295 - 297
  • [28] Knowledge-Enhanced Language Models Are Not Bias-Proof: Situated Knowledge and Epistemic Injustice in AI
    Kraft, Angelie
    Soulier, Eloise
    PROCEEDINGS OF THE 2024 ACM CONFERENCE ON FAIRNESS, ACCOUNTABILITY, AND TRANSPARENCY, ACM FACCT 2024, 2024, : 1433 - 1445
  • [29] CoPL: Contextual Prompt Learning for Vision-Language Understanding
    Goswami, Koustava
    Karanam, Srikrishna
    Udhayanan, Prateksha
    Joseph, K. J.
    Srinivasan, Balaji Vasan
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 16, 2024, : 18090 - 18098
  • [30] Semantic vector learning for natural language understanding
    Jung, Sangkeun
    COMPUTER SPEECH AND LANGUAGE, 2019, 56 : 130 - 145