RECURRENT NEURAL NETWORK LANGUAGE MODEL TRAINING WITH NOISE CONTRASTIVE ESTIMATION FOR SPEECH RECOGNITION

被引:0
|
作者
Chen, X. [1 ]
Liu, X. [1 ]
Gales, M. J. E. [1 ]
Woodland, P. C. [1 ]
机构
[1] Univ Cambridge, Engn Dept, Trumpington St, Cambridge CB2 1PZ, England
基金
英国工程与自然科学研究理事会;
关键词
language model; recurrent neural network; GPU; noise contrastive estimation; speech recognition;
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
In recent years recurrent neural network language models (RNNLMs) have been successfully applied to a range of tasks including speech recognition. However, an important issue that limits the quantity of data used, and their possible application areas, is the computational cost in training. A significant part of this cost is associated with the softmax function at the output layer, as this requires a normalization term to be explicitly calculated. This impacts both the training and testing speed, especially when a large output vocabulary is used. To address this problem, noise contrastive estimation (NCE) is explored in RNNLM training. NCE does not require the above normalization during both training and testing. It is insensitive to the output layer size. On a large vocabulary conversational telephone speech recognition task, a doubling in training speed on a GPU and a 56 times speed up in test time evaluation on a CPU were obtained.
引用
收藏
页码:5411 / 5415
页数:5
相关论文
共 50 条
  • [21] Speech Emotion Recognition Based on a Recurrent Neural Network Classification Model
    Fonnegra, Ruben D.
    Diaz, Gloria M.
    ADVANCES IN COMPUTER ENTERTAINMENT TECHNOLOGY, ACE 2017, 2018, 10714 : 882 - 892
  • [22] Speech Recognition Model for Assamese Language Using Deep Neural Network
    Singh, Moirangthem Tiken
    Barman, Partha Pratim
    Gogoi, Rupjyoti
    2018 INTERNATIONAL CONFERENCE ON RECENT INNOVATIONS IN ELECTRICAL, ELECTRONICS & COMMUNICATION ENGINEERING (ICRIEECE 2018), 2018, : 2722 - 2727
  • [23] REAL-TIME ONE-PASS DECODING WITH RECURRENT NEURAL NETWORK LANGUAGE MODEL FOR SPEECH RECOGNITION
    Hori, Takaaki
    Kubo, Yotaro
    Nakamura, Atsushi
    2014 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2014,
  • [24] RECURRENT NEURAL NETWORK LANGUAGE MODEL TRAINING USING NATURAL GRADIENT
    Yu, Jianwei
    Lam, Max. W. Y.
    Chen, Xie
    Hu, Shoukang
    Liu, Songxiang
    Wu, Xixin
    Liu, Xunying
    Meng, Helen
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 7260 - 7264
  • [25] Recurrent Neural Network Based Language Model Adaptation for Accent Mandarin Speech
    Ni, Hao
    Yi, Jiangyan
    Wen, Zhengqi
    Tao, Jianhua
    PATTERN RECOGNITION (CCPR 2016), PT II, 2016, 663 : 607 - 617
  • [26] DEEP RECURRENT REGULARIZATION NEURAL NETWORK FOR SPEECH RECOGNITION
    Chien, Jen-Tzung
    Lu, Tsai-Wei
    2015 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING (ICASSP), 2015, : 4560 - 4564
  • [27] Implementation of an autoassociative Recurrent Neural Network for speech recognition
    Cocchiglia, A
    Paplinski, A
    IEEE TENCON'97 - IEEE REGIONAL 10 ANNUAL CONFERENCE, PROCEEDINGS, VOLS 1 AND 2: SPEECH AND IMAGE TECHNOLOGIES FOR COMPUTING AND TELECOMMUNICATIONS, 1997, : 245 - 248
  • [28] MINIMUM WORD ERROR TRAINING OF LONG SHORT-TERM MEMORY RECURRENT NEURAL NETWORK LANGUAGE MODELS FOR SPEECH RECOGNITION
    Hori, Takaaki
    Hori, Chiori
    Watanabe, Shinji
    Hershey, John R.
    2016 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING PROCEEDINGS, 2016, : 5990 - 5994
  • [29] Time Delay Recurrent Neural Network for Speech Recognition
    Liu, Boji
    Zhang, Weibin
    Xu, Xiangming
    Chen, Dongpeng
    2019 3RD INTERNATIONAL CONFERENCE ON MACHINE VISION AND INFORMATION TECHNOLOGY (CMVIT 2019), 2019, 1229
  • [30] Recurrent Neural Network-based Language Modeling for an Automatic Russian Speech Recognition System
    Kipyatkova, Irina
    Karpov, Alexey
    2015 ARTIFICIAL INTELLIGENCE AND NATURAL LANGUAGE AND INFORMATION EXTRACTION, SOCIAL MEDIA AND WEB SEARCH FRUCT CONFERENCE (AINL-ISMW FRUCT), 2015, : 33 - 38