INVESTIGATION ON LOG-LINEAR INTERPOLATION OF MULTI-DOMAIN NEURAL NETWORK LANGUAGE MODEL

被引:0
|
作者
Tueske, Zoltan [1 ]
Irie, Kazuki [1 ]
Schlueter, Ralf [1 ]
Ney, Hermann [1 ]
机构
[1] Rhein Westfal TH Aachen, Dept Comp Sci, Human Language Technol & Pattern Recognit, D-52056 Aachen, Germany
关键词
multi-domain; language modeling; deep feed-forward network; LM adaptation; log-linear; interpolation;
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Inspired by the success of multi-task training in acoustic modeling, this paper investigates a new architecture for a multi-domain neural network based language model (NNLM). The proposed model has several shared hidden layers and domain-specific output layers. As will be shown, the log-linear interpolation of the multi-domain outputs and the optimization of interpolation weights fit naturally in the framework of NNLM. The resulting model can be expressed as a single NNLM. As an initial study of such an architecture, this paper focuses on deep feed-forward neural networks (DNNs). We also re-investigate the potential of long context up to 30-grams, and depth up to 5 hidden layers in DNN-LM. Our final feed-forward multi-domain NNLM is trained on 3.1B running words across 11 domains for English broadcast news and conversations large vocabulary continuous speech recognition task. After log-linear interpolation and fine-tuning, we measured improvements in terms of perplexity and word error rate over the models trained on 50M running words of in-domain news resources. The final multi-domain feed-forward LM outperformed our previous best LSTM-RNN LM trained on the 50M in-domain corpus, even after linear interpolation with large count models.
引用
收藏
页码:6005 / 6009
页数:5
相关论文
共 50 条
  • [1] Multi-domain Neural Network Language Model
    Alumae, Tanel
    [J]. 14TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2013), VOLS 1-5, 2013, : 2181 - 2185
  • [2] Normalized Log-Linear Interpolation of Backoff Language Models is Efficient
    Heafield, Kenneth
    Geigle, Chase
    Massung, Sean
    Schwartz, Lane
    [J]. PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, 2016, : 876 - 886
  • [3] Multi-Domain Recurrent Neural Network Language Model for Medical Speech Recognition
    Tilk, Ottokar
    Alumaee, Tanel
    [J]. HUMAN LANGUAGE TECHNOLOGIES - THE BALTIC PERSPECTIVE, BALTIC HLT 2014, 2014, 268 : 149 - +
  • [4] LOG-LINEAR MODEL
    SATO, Y
    [J]. SOCIOLOGICAL THEORY AND METHODS, 1995, 10 (01) : 77 - 90
  • [5] Multi-Domain Neural Network Recommender
    Yi, Baolin
    Zhao, Shuting
    Shen, Xiaoxuan
    Zhang, Li
    [J]. 2018 IEEE INTERNATIONAL CONFERENCE ON ELECTRONICS AND COMMUNICATION ENGINEERING (ICECE 2018), 2018, : 41 - 45
  • [7] Multi-domain Attention Fusion Network For Language Recognition
    Ju M.
    Xu Y.
    Ke D.
    Su K.
    [J]. SN Computer Science, 4 (1)
  • [8] MODEL SELECTION FOR LOG-LINEAR MODELS
    BAI, ZD
    KRISHNAIAH, PR
    SAMBAMOORTHI, N
    ZHAO, LC
    [J]. SANKHYA-THE INDIAN JOURNAL OF STATISTICS SERIES B, 1992, 54 : 200 - 219
  • [9] A multi-level log-linear model of childhood leukaemia mortality
    Langford, Ian H.
    [J]. HEALTH & PLACE, 1995, 1 (02) : 113 - 119
  • [10] PRICE EXPECTATION - MULTI-VARIED LOG-LINEAR MODEL OF PROBABILITY
    KONIG, H
    [J]. KYKLOS, 1979, 32 (1-2) : 380 - 391