ADAPTABLE MULTI-DOMAIN LANGUAGE MODEL FOR TRANSFORMER ASR

被引:4
|
作者
Lee, Taewoo [1 ]
Lee, Min-Joong [2 ]
Kang, Tae Gyoon [2 ]
Jung, Seokyeoung [1 ]
Kwon, Minseok [1 ]
Hong, Yeona [1 ]
Lee, Jungin [1 ]
Woo, Kyoung-Gu [1 ]
Kim, Ho-Gyeong [2 ]
Jeong, Jiseung [2 ]
Lee, Jihyun [2 ]
Lee, Hosik [2 ]
Choi, Young Sang [2 ]
机构
[1] Samsung Elect, AI R&D Grp, Suwon Shi, South Korea
[2] Samsung Elect, Samsung Adv Inst Technol, Suwon Shi, South Korea
关键词
End-to-end (E2E) automatic speech recognition (ASR); language model (LM); multi-domain adaptation;
D O I
10.1109/ICASSP39728.2021.9413475
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
We propose an adapter based multi-domain Transformer based language model (LM) for Transformer ASR. The model consists of a big size common LM and small size adapters. The model can perform multi-domain adaptation with only the small size adapters and its related layers. The proposed model can reuse the full fine-tuned LM which is fine-tuned using all layers of an original model. The proposed LM can be expanded to new domains by adding about 2% of parameters for a first domain and 13% parameters for after second domain. The proposed model is also effective in reducing the model maintenance cost because it is possible to omit the costly and time-consuming common LM pre-training process. Using proposed adapter based approach, we observed that a general LM with adapter can outperform a dedicated music domain LM in terms of word error rate (WER).
引用
收藏
页码:7358 / 7362
页数:5
相关论文
共 50 条
  • [41] Path computation in multi-layer multi-domain networks: A language theoretic approach
    Lamali, Mohamed Lamine
    Pouyllau, Helia
    Barth, Dominique
    COMPUTER COMMUNICATIONS, 2013, 36 (05) : 589 - 599
  • [42] Improved Hybrid Streaming ASR with Transformer Language Models
    Baquero-Arnal, Pau
    Jorge, Javier
    Gimenez, Adria
    Albert Silvestre-Cerda, Joan
    Iranzo-Sanchez, Javier
    Sanchis, Albert
    Civera, Jorge
    Juan, Alfons
    INTERSPEECH 2020, 2020, : 2127 - 2131
  • [43] Semi-supervised single- and multi-domain regression with multi-domain training
    Michaeli, Tomer
    Eldar, Yonina C.
    Sapiro, Guillermo
    INFORMATION AND INFERENCE-A JOURNAL OF THE IMA, 2012, 1 (01) : 68 - 97
  • [44] Multi-domain gate and interactive dual attention for multi-domain dialogue state tracking
    Jia, Xu
    Zhang, Ruochen
    Peng, Min
    KNOWLEDGE-BASED SYSTEMS, 2024, 286
  • [45] Multi-domain gate and interactive dual attention for multi-domain dialogue state tracking
    Jia, Xu
    Zhang, Ruochen
    Peng, Min
    Knowledge-Based Systems, 2024, 286
  • [46] Entity-centric multi-domain transformer for improving generalization in fake news detection
    Bazmi, Parisa
    Asadpour, Masoud
    Shakery, Azadeh
    Maazallahi, Abbas
    INFORMATION PROCESSING & MANAGEMENT, 2024, 61 (05)
  • [47] Multi-domain integrative Swin transformer network for sparse-view tomographic reconstruction
    Pan, Jiayi
    Zhang, Heye
    Wu, Weifei
    Gao, Zhifan
    Wu, Weiwen
    PATTERNS, 2022, 3 (06):
  • [48] Multi-domain Spoken Language Understanding Using Domain- and Task-aware Parameterization
    Qin, Libo
    Wei, Fuxuan
    Ni, Minheng
    Zhang, Yue
    Che, Wanxiang
    Li, Yangming
    Liu, Ting
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2022, 21 (04)
  • [49] A novel multi-domain machine reading comprehension model with domain interference mitigation
    Zhou, Chulun
    Wang, Zhihao
    He, Shaojie
    Zhang, Haiying
    Su, Jinsong
    NEUROCOMPUTING, 2022, 500 : 791 - 798
  • [50] Generation of nearshore bars by multi-domain hybrid numerical model
    Lee, CE
    Kim, MH
    Edge, BL
    JOURNAL OF COASTAL RESEARCH, 1999, 15 (04) : 892 - 901