LEARNING NEURAL TRANS-DIMENSIONAL RANDOM FIELD LANGUAGE MODELS WITH NOISE-CONTRASTIVE ESTIMATION

被引:0
|
作者
Wang, Bin [1 ]
Ou, Zhijian [1 ]
机构
[1] Tsinghua Univ, Speech Proc & Machine Intelligence SPMI Lab, Beijing, Peoples R China
关键词
Language Model; Random Field; Speech Recognition; Noise-contrastive Estimation;
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Trans-dimensional random field language models (TRF LMs) where sentences are modeled as a collection of random fields, have shown close performance with LSTM LMs in speech recognition and are computationally more efficient in inference. However, the training efficiency of neural TRF LMs is not satisfactory, which limits the scalability of TRF LMs on large training corpus. In this paper, several techniques on both model formulation and parameter estimation are proposed to improve the training efficiency and the performance of neural TRF LMs. First, TRFs are reformulated in the form of exponential tilting of a reference distribution. Second, noise-contrastive estimation (NCE) is introduced to jointly estimate the model parameters and normalization constants. Third, we extend the neural TRF LMs by marrying the deep convolutional neural network (CNN) and the bidirectional LSTM into the potential function to extract the deep hierarchical features and bidirectionally sequential features. Utilizing all the above techniques enables the successful and efficient training of neural TRF LMs on a 40x larger training set with only 1/3 training time and further reduces the WER with relative reduction of 4.7% on top of a strong LSTM LM baseline.
引用
收藏
页码:6134 / 6138
页数:5
相关论文
共 20 条
  • [1] IMPROVED TRAINING OF NEURAL TRANS-DIMENSIONAL RANDOM FIELD LANGUAGE MODELS WITH DYNAMIC NOISE-CONTRASTIVE ESTIMATION
    Wang, Bin
    Ou, Zhijian
    2018 IEEE WORKSHOP ON SPOKEN LANGUAGE TECHNOLOGY (SLT 2018), 2018, : 70 - 76
  • [2] LANGUAGE MODELING WITH NEURAL TRANS-DIMENSIONAL RANDOM FIELDS
    Wang, Bin
    Ou, Zhijian
    2017 IEEE AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING WORKSHOP (ASRU), 2017, : 294 - 300
  • [3] Conditional Noise-Contrastive Estimation of Unnormalised Models
    Ceylan, Ciwan
    Gutmann, Michael U.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [4] INTEGRATING DISCRETE AND NEURAL FEATURES VIA MIXED-FEATURE TRANS-DIMENSIONAL RANDOM FIELD LANGUAGE MODELS
    Gao, Silin
    Ou, Zhijian
    Yang, Wei
    Xu, Huifang
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 6169 - 6173
  • [5] Learning Trans-Dimensional Random Fields with Applications to Language Modeling
    Wang, Bin
    Ou, Zhijian
    Tan, Zhiqiang
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (04) : 876 - 890
  • [6] Trans-dimensional Random Fields for Language Modeling
    Wang, Bin
    Ou, Zhijian
    Tan, Zhiqiang
    PROCEEDINGS OF THE 53RD ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 7TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1, 2015, : 785 - 794
  • [7] Trans-dimensional random fields for language modeling
    Department of Electronic Engineering, Tsinghua University, Beijing
    100084, China
    不详
    NJ
    08854, United States
    ACL-IJCNLP - Annu. Meet. Assoc. Comput. Linguist. Int. Jt. Conf. Nat. Lang. Process. Asian Fed. Nat. Lang. Process., Proc. Conf., (785-794):
  • [8] Noise-Contrastive Estimation for Answer Selection with Deep Neural Networks
    Rao, Jinfeng
    He, Hua
    Lin, Jimmy
    CIKM'16: PROCEEDINGS OF THE 2016 ACM CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, 2016, : 1913 - 1916
  • [9] Noise-Contrastive Estimation of Unnormalized Statistical Models, with Applications to Natural Image Statistics
    Gutmann, Michael U.
    Hyvarinen, Aapo
    JOURNAL OF MACHINE LEARNING RESEARCH, 2012, 13 : 307 - 361
  • [10] Noise-Contrastive Estimation Based on Relative Neighbour Sampling for Unsupervised Image Embedding Learning
    Lv, Haoyu
    Cao, Jiangzhong
    Dai, Qingyun
    Liang, Guanshu
    Zhao, Qi
    2019 4TH IEEE INTERNATIONAL CONFERENCE ON BIG DATA ANALYTICS (ICBDA 2019), 2019, : 308 - 311