Knowledge enhancement BERT based on domain dictionary mask

被引:0
|
作者
Cao, Xianglin [1 ]
Xiao, Hong [1 ]
Jiang, Wenchao [1 ]
机构
[1] Guangdong Univ Technol, Sch Comp Sci & Technol, Guangzhou, Peoples R China
关键词
Intelligent customer service; dictionary mask; BERT; data preprocessing;
D O I
10.3233/JHS-222013
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Semantic matching is one of the critical technologies for intelligent customer service. Since Bidirectional Encoder Representations from Transformers (BERT) is proposed, fine-tuning on a large-scale pre-training language model becomes a general method to implement text semantic matching. However, in practical application, the accuracy of the BERT model is limited by the quantity of pre-training corpus and proper nouns in the target domain. An enhancement method for knowledge based on domain dictionary to mask input is proposed to solve the problem. Firstly, for modul input, we use keyword matching to recognize and mask the word in domain. Secondly, using self-supervised learning to inject knowledge of the target domain into the BERT model. Thirdly, we fine-tune the BERT model with public datasets LCQMC and BQboost. Finally, we test the model's performance with a financial company's user data. The experimental results show that after using our method and BQboost, accuracy increases by 12.12% on average in practical applications.
引用
收藏
页码:121 / 128
页数:8
相关论文
共 50 条
  • [1] Domain Terminology Knowledge Graph Completion Method Based on Bert
    Huang, Weichun
    Yang, Jian
    Xiao, Gang
    Hu, Xinyu
    2023 11TH INTERNATIONAL CONFERENCE ON INFORMATION TECHNOLOGY: IOT AND SMART CITY, ITIOTSC 2023, 2023, : 1 - 4
  • [2] How Can the [MASK] Know? The Sources and Limitations of Knowledge in BERT
    Podkorytov, Maksim
    Bis, Daniel
    Liu, Xiuwen
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [3] Knowledge distillation for BERT unsupervised domain adaptation
    Ryu, Minho
    Lee, Geonseok
    Lee, Kichun
    KNOWLEDGE AND INFORMATION SYSTEMS, 2022, 64 (11) : 3113 - 3128
  • [4] Knowledge distillation for BERT unsupervised domain adaptation
    Minho Ryu
    Geonseok Lee
    Kichun Lee
    Knowledge and Information Systems, 2022, 64 : 3113 - 3128
  • [5] Improving BERT-Based Text Classification With Auxiliary Sentence and Domain Knowledge
    Yu, Shanshan
    Su, Jindian
    Luo, Da
    IEEE ACCESS, 2019, 7 : 176600 - 176612
  • [6] Knowledge Distillation with Feature Enhancement Mask
    Xiao, Yue
    Wang, Longye
    Li, Wentao
    Zeng, Xiaoli
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT VIII, 2023, 14261 : 432 - 443
  • [7] Exploitation of domain knowledge by knowledge-based database design tools: The dictionary approach
    Noah, SA
    Lloyd-Williams, M
    KNOWLEDGE-BASED SOFTWARE ENGINEERING, 1998, 48 : 250 - 257
  • [8] Colour Retinal Image Enhancement based on Domain Knowledge
    Joshi, Gopal Datt
    Sivaswamy, Jayanthi
    SIXTH INDIAN CONFERENCE ON COMPUTER VISION, GRAPHICS & IMAGE PROCESSING ICVGIP 2008, 2008, : 591 - 598
  • [9] Chinese Named Entity Recognition Method Based on Dictionary Semantic Knowledge Enhancement
    Wang, Tianbin
    Huang, Ruiyang
    Hu, Nan
    Wang, Huansha
    Chu, Guanghan
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2023, E106D (05) : 1010 - 1017
  • [10] C-BERT: A Mongolian reverse dictionary based on fused lexical semantic clustering and BERT
    Wang, Amuguleng
    Qi, Yilagui
    Baiyila, Dahu
    ALEXANDRIA ENGINEERING JOURNAL, 2025, 111 : 385 - 395