Joint Unsupervised Adaptation of N-gram and RNN Language Models via LDA-based Hybrid Mixture Modeling

被引:0
|
作者
Masumura, Ryo [1 ]
Asami, Taichi [1 ]
Masataki, Hirokazu [1 ]
Aono, Yushi [1 ]
机构
[1] NTT Corp, NTT Media Intelligence Labs, Tokyo, Japan
关键词
D O I
暂无
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
This paper reports an initial study of unsupervised adaptation that assumes simultaneous use of both n-gram and recurrent neural network (RNN) language models (LMs) in automatic speech recognition (ASR). It is known that a combination of n-grams and RNN LMs is a more effective approach to ASR than using each of them singly. However, unsupervised adaptation methods that simultaneously adapt both n-grams and RNN LMs have not been presented while various unsupervised adaptation methods specific to either n-gram LMs or RNN LMs have been examined. In order to handle different LMs in a unified unsupervised adaptation framework, our key idea is to introduce mixture modeling for both n-gram LMs and RNN LMs. The mixture modeling can simultaneously handle multiple LMs and unsupervised adaptation can be easily accomplished merely by adjusting their mixture weights using a recognition hypothesis of an input speech. This paper proposes joint unsupervised adaptation achieved by a hybrid mixture modeling using both n-gram mixture models and RNN mixture models. We present latent Dirichlet allocation based hybrid mixture modeling for effective topic adaptation. Our experiments in lecture ASR tasks show the effectiveness of joint unsupervised adaptation. We also reveal performance in which only one n-gram or RNN LM is adapted.
引用
下载
收藏
页码:1538 / 1541
页数:4
相关论文
共 33 条
  • [21] A Spectral Algorithm for Learning Class-Based n-gram Models of Natural Language
    Stratos, Karl
    Kim, Do-kyum
    Collins, Michael
    Hsu, Daniel
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2014, : 762 - 771
  • [23] Taming Pre-trained Language Models with N-gram Representations for Low-Resource Domain Adaptation
    Diao, Shizhe
    Xu, Ruijia
    Su, Hongjin
    Jiang, Yilei
    Song, Yan
    Zhang, Tong
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, 2021, : 3336 - 3349
  • [24] N-gram Adaptation Using Dirichlet Class Language Model Based on Part-of-Speech for Speech Recognition
    Hatami, Ali
    Akbari, Ahmad
    Nasersharif, Babak
    2013 21ST IRANIAN CONFERENCE ON ELECTRICAL ENGINEERING (ICEE), 2013,
  • [25] Unsupervised language model adaptation via topic modeling based on named entity hypotheses
    Liu, Yang
    Liu, Feifan
    2008 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, VOLS 1-12, 2008, : 4921 - 4924
  • [26] INVESTIGATION OF BACK-OFF BASED INTERPOLATION BETWEEN RECURRENT NEURAL NETWORK AND N-GRAM LANGUAGE MODELS
    Chen, X.
    Liu, X.
    Gales, M. J. F.
    Woodland, P. C.
    2015 IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING (ASRU), 2015, : 181 - 186
  • [27] Interpolation of n-gram and mutual-information based trigger pair language models for Mandarin speech recognition
    Zhou, GD
    Lua, KT
    COMPUTER SPEECH AND LANGUAGE, 1999, 13 (02): : 125 - 141
  • [28] Interpolation of n-gram and mutual-information based trigger pair language models for Mandarin speech recognition
    Department of Computer Science, School of Computing, National University of Singapore, Lower Kent Ridge Road, Singapore 119260, Singapore
    Comput Speech Lang, 2 (125-141):
  • [29] N-Gram Based Sarcasm Detection for News and Social Media Text Using Hybrid Deep Learning Models
    Thaokar C.
    Rout J.K.
    Rout M.
    Ray N.K.
    SN Computer Science, 5 (1)
  • [30] Unsupervised acquisition of idiomatic units of symbolic natural language: An n-gram frequency-based approach for the chunking of news articles and tweets
    Borrelli, Dario
    Gongora Svartzman, Gabriela
    Lipizzi, Carlo
    PLOS ONE, 2020, 15 (06):