ADAPTABLE MULTI-DOMAIN LANGUAGE MODEL FOR TRANSFORMER ASR

被引:4
|
作者
Lee, Taewoo [1 ]
Lee, Min-Joong [2 ]
Kang, Tae Gyoon [2 ]
Jung, Seokyeoung [1 ]
Kwon, Minseok [1 ]
Hong, Yeona [1 ]
Lee, Jungin [1 ]
Woo, Kyoung-Gu [1 ]
Kim, Ho-Gyeong [2 ]
Jeong, Jiseung [2 ]
Lee, Jihyun [2 ]
Lee, Hosik [2 ]
Choi, Young Sang [2 ]
机构
[1] Samsung Elect, AI R&D Grp, Suwon Shi, South Korea
[2] Samsung Elect, Samsung Adv Inst Technol, Suwon Shi, South Korea
关键词
End-to-end (E2E) automatic speech recognition (ASR); language model (LM); multi-domain adaptation;
D O I
10.1109/ICASSP39728.2021.9413475
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
We propose an adapter based multi-domain Transformer based language model (LM) for Transformer ASR. The model consists of a big size common LM and small size adapters. The model can perform multi-domain adaptation with only the small size adapters and its related layers. The proposed model can reuse the full fine-tuned LM which is fine-tuned using all layers of an original model. The proposed LM can be expanded to new domains by adding about 2% of parameters for a first domain and 13% parameters for after second domain. The proposed model is also effective in reducing the model maintenance cost because it is possible to omit the costly and time-consuming common LM pre-training process. Using proposed adapter based approach, we observed that a general LM with adapter can outperform a dedicated music domain LM in terms of word error rate (WER).
引用
收藏
页码:7358 / 7362
页数:5
相关论文
共 50 条
  • [1] Multi-domain Neural Network Language Model
    Alumae, Tanel
    14TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2013), VOLS 1-5, 2013, : 2181 - 2185
  • [2] Unified Modeling of Multi-Domain Multi-Device ASR Systems
    Mitra, Soumyajit
    Ray, Swayambhu Nath
    Padil, Bharat
    Bilgi, Raghavendra
    Arsiker, Harish
    Ghosh, Shalini
    Srinivasamurthy, Ajay
    Garimella, Sri
    TEXT, SPEECH, AND DIALOGUE, TSD 2023, 2023, 14102 : 283 - 292
  • [3] Factorized Transformer for Multi-Domain Neural Machine Translation
    Deng, Yongchao
    Yu, Hongfei
    Yu, Heng
    Duan, Xiangyu
    Luo, Weihua
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 4221 - 4230
  • [4] Large Scale, Multi-domain Language Identification
    Jauhiainen, Tommi
    Zampieri, Marcos
    Baldwin, Timothy
    Lindén, Krister
    Synthesis Lectures on Human Language Technologies, 2024, Part F2039 : 117 - 135
  • [5] Continually Tuning a Large Language Model for Multi-domain Radiology Report Generation
    Sun, Yihua
    Khor, Hee Guan
    Wang, Yuanzheng
    Wang, Zhuhao
    Zhao, Hongliang
    Zhang, Yu
    Ma, Longfei
    Zheng, Zhuozhao
    Liao, Hongen
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2024, PT V, 2024, 15005 : 177 - 187
  • [6] Multi-Domain Recurrent Neural Network Language Model for Medical Speech Recognition
    Tilk, Ottokar
    Alumaee, Tanel
    HUMAN LANGUAGE TECHNOLOGIES - THE BALTIC PERSPECTIVE, BALTIC HLT 2014, 2014, 268 : 149 - +
  • [7] A Multi-Domain Role Activation Model
    Abreu, Vilmar
    Santin, Altair O.
    Viegas, Eduardo K.
    Stihler, Maicon
    2017 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC), 2017,
  • [8] Domain attention model for multi-domain sentiment classification
    Yuan, Zhigang
    Wu, Sixing
    Wu, Fangzhao
    Liu, Junxin
    Huang, Yongfeng
    KNOWLEDGE-BASED SYSTEMS, 2018, 155 : 1 - 10
  • [9] Multi-domain Attention Fusion Network For Language Recognition
    Ju M.
    Xu Y.
    Ke D.
    Su K.
    SN Computer Science, 4 (1)
  • [10] ClippyScript: A Programming Language for Multi-Domain Dialogue Systems
    Seide, Frank
    McDirmid, Sean
    13TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2012 (INTERSPEECH 2012), VOLS 1-3, 2012, : 242 - 245