EFFICIENT TEXT ANALYSIS WITH PRE-TRAINED NEURAL NETWORK MODELS

被引:1
|
作者
Cui, Jia [1 ]
Lu, Heng [1 ,3 ]
Wang, Wenjie [2 ]
Kang, Shiyin [1 ,4 ]
He, Liqiang [1 ]
Li, Guangzhi [1 ]
Yu, Dong [1 ]
机构
[1] Tencent AI Lab, Seattle, WA 98004 USA
[2] Emory Univ, Atlanta, GA 30322 USA
[3] Ximalaya Inc, Shanghai, Peoples R China
[4] Huya Inc, Guangzhou, Peoples R China
关键词
Text analysis; TTS frontend; G2P; text normalization; punctuation; weakly supervised learning; phrase-based attention;
D O I
10.1109/SLT54892.2023.10022565
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper investigates the application of pre-trained BERT model in three classic text analysis tasks: Chinese grapheme-tophoneme(G2P), text normalization(TN) and sentence punctuation annotation. Even though the full-sized BERT has prominent modeling power, there are two challenges for it in real applications: the requirement for annotated training data and the considerable computational cost. In this paper, we propose BERT-based low-latency solutions. To collect sufficient training corpus for G2P, we transfer knowledge from existing rule-based system to BERT through a large amount of unlabeled corpus. The new model could convert all characters directly from raw texts with higher accuracy. We also propose a hybrid two-stage text normalization pipeline which reduces the sentence error rate by 25% compared to the rule-based system. We offer both supervised and weakly supervised versions and find that the latter has only 1% accuracy drop from the former.
引用
收藏
页码:671 / 676
页数:6
相关论文
共 50 条
  • [41] Bi-tuning: Efficient Transfer from Pre-trained Models
    Zhong, Jincheng
    Ma, Haoyu
    Wang, Ximei
    Kou, Zhi
    Long, Mingsheng
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, ECML PKDD 2023, PT V, 2023, 14173 : 357 - 373
  • [42] EFFICIENT UTILIZATION OF LARGE PRE-TRAINED MODELS FOR LOW RESOURCE ASR
    Vieting, Peter
    Luescher, Christoph
    Dierkes, Julian
    Schlueter, Ralf
    Ney, Hermann
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING WORKSHOPS, ICASSPW, 2023,
  • [43] Pre-Trained Language Models and Their Applications
    Wang, Haifeng
    Li, Jiwei
    Wu, Hua
    Hovy, Eduard
    Sun, Yu
    [J]. ENGINEERING, 2023, 25 : 51 - 65
  • [44] Neural Transfer Learning For Vietnamese Sentiment Analysis Using Pre-trained Contextual Language Models
    An Pha Le
    Tran Vu Pham
    Thanh-Van Le
    Huynh, Duy, V
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLIED NETWORK TECHNOLOGIES (ICMLANT II), 2021, : 84 - 88
  • [45] Vulnerability Analysis of Continuous Prompts for Pre-trained Language Models
    Li, Zhicheng
    Shi, Yundi
    Sheng, Xuan
    Yin, Changchun
    Zhou, Lu
    Li, Piji
    [J]. ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT IX, 2023, 14262 : 508 - 519
  • [46] Transformer over Pre-trained Transformer for Neural Text Segmentation with Enhanced Topic Coherence
    Lo, Kelvin
    Jin, Yuan
    Tan, Weicong
    Liu, Ming
    Du, Lan
    Buntine, Wray
    [J]. FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 3334 - 3340
  • [47] Difference between Multi-modal vs. Text Pre-trained Models in Embedding Text
    Sun Y.
    Cheng X.
    Song R.
    Che W.
    Lu Z.
    Wen J.
    [J]. Beijing Daxue Xuebao (Ziran Kexue Ban)/Acta Scientiarum Naturalium Universitatis Pekinensis, 2023, 59 (01): : 48 - 56
  • [48] Neural machine translation of clinical text: an empirical investigation into multilingual pre-trained language models and transfer-learning
    Han, Lifeng
    Gladkoff, Serge
    Erofeev, Gleb
    Sorokina, Irina
    Galiano, Betty
    Nenadic, Goran
    [J]. FRONTIERS IN DIGITAL HEALTH, 2024, 6
  • [49] How Different are Pre-trained Transformers for Text Ranking?
    Rau, David
    Kamps, Jaap
    [J]. ADVANCES IN INFORMATION RETRIEVAL, PT II, 2022, 13186 : 207 - 214
  • [50] CUE: An Uncertainty Interpretation Framework for Text Classifiers Built on Pre-Trained Language Models
    Li, Jiazheng
    Sun, Zhaoyue
    Liang, Bin
    Gui, Lin
    He, Yulan
    [J]. UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2023, 216 : 1253 - 1262