Make BERT-based Chinese Spelling Check Model Enhanced by Layerwise Attention and Gaussian

被引:0
|
作者
Cao, Yongchang [1 ]
He, Liang [1 ]
Wu, Zhen [1 ]
Dai, Xinyu [1 ]
机构
[1] Nanjing Univ, Natl Key Lab Novel Software Technol, Nanjing, Peoples R China
关键词
D O I
10.1109/IJCNN54540.2023.10191265
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
BERT-based models have shown a remarkable ability in the Chinese Spelling Check (CSC) task recently. However, traditional BERT-based methods still suffer from two limitations. First, although previous works have identified that explicit prior knowledge like Part-Of-Speech (POS) tagging can benefit in the CSC task, they neglected the fact that spelling errors inherent in CSC data can lead to incorrect tags and therefore mislead models. Additionally, they ignored the correlation between the implicit hierarchical information encoded by BERT's intermediate layers and different linguistic phenomena. This results in sub-optimal accuracy. To alleviate the above two issues, we design a heterogeneous knowledge-infused framework to strengthen BERT-based CSC models. To incorporate explicit POS knowledge, we utilize an auxiliary task strategy driven by Gaussian mixture model. Meanwhile, to incorporate implicit hierarchical linguistic knowledge within the encoder, we propose a novel form of n-gram-based layerwise self-attention to generate a multilayer representation. Experimental results show that our proposed framework yields a stable performance boost over four strong baseline models and outperforms the previous state-of-the-art methods on two datasets.
引用
收藏
页数:9
相关论文
共 50 条
  • [1] Bert-Based Chinese Medical Keyphrase Extraction Model Enhanced with External Features
    Ding, Liangping
    Zhang, Zhixiong
    Zhao, Yang
    [J]. TOWARDS OPEN AND TRUSTWORTHY DIGITAL SOCIETIES, ICADL 2021, 2021, 13133 : 167 - 176
  • [2] A BERT-Based Two-Stage Model for Chinese Chengyu Recommendation
    Tan, Minghuan
    Jiang, Jing
    Dai, Bing Tian
    [J]. ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2021, 20 (06)
  • [3] BEDSpell: Spelling Error Correction Using BERT-Based Masked Language Model and Edit Distance
    Tohidian, Fatemeh
    Kashiri, Amin
    Lotfi, Fariba
    [J]. SERVICE-ORIENTED COMPUTING - ICSOC 2022 WORKSHOPS, 2023, 13821 : 3 - 14
  • [4] A BERT-based Idiom Detection Model
    Gamage, Gihan
    De Silva, Daswin
    Adikari, Achini
    Alahakoon, Damminda
    [J]. 2022 15TH INTERNATIONAL CONFERENCE ON HUMAN SYSTEM INTERACTION (HSI), 2022,
  • [5] A Hybrid Model for Chinese Spelling Check
    Zhao, Hai
    Cai, Deng
    Xin, Yang
    Wang, Yuzhu
    Jia, Zhongye
    [J]. ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2017, 16 (03)
  • [6] Chinese Grammatical Correction Using BERT-based Pre-trained Model
    Wang, Hongfei
    Kurosawa, Michiki
    Katsumatat, Satoru
    Komachi, Mamoru
    [J]. 1ST CONFERENCE OF THE ASIA-PACIFIC CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 10TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (AACL-IJCNLP 2020), 2020, : 163 - 168
  • [7] BERT-Based Chinese Relation Extraction for Public Security
    Hou, Jiaqi
    Li, Xin
    Yao, Haipeng
    Sun, Haichun
    Mai, Tianle
    Zhu, Rongchen
    [J]. IEEE ACCESS, 2020, 8 : 132367 - 132375
  • [8] A BERT-based interactive attention network for aspect sentiment analysis
    Yang, Yu-Ting
    Feng, Lin
    Dai, Lei-Chao
    [J]. Journal of Computers (Taiwan), 2021, 32 (03) : 30 - 42
  • [9] A BERT-Based Semantic Enhanced Model for COVID-19 Fake News Detection
    Yin, Hui
    Liu, Xiao
    Wu, Yutao
    Aria, Hilya Mudrika
    Mohawesh, Rami
    [J]. WEB AND BIG DATA, PT I, APWEB-WAIM 2023, 2024, 14331 : 1 - 15
  • [10] Assessing the use of attention weights to interpret BERT-based stance classification
    Cordova Saenz, Carlos Abel
    Becker, Karin
    [J]. 2021 IEEE/WIC/ACM INTERNATIONAL CONFERENCE ON WEB INTELLIGENCE AND INTELLIGENT AGENT TECHNOLOGY (WI-IAT 2021), 2021, : 194 - 201