Efficient training of large neural networks for language modeling

被引:0
|
作者
Schwenk, H [1 ]
机构
[1] CNRS, LIMSI, F-91403 Orsay, France
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently there has been increasing interest in using neural networks for language modeling. In contrast to the well known backoff n-gram language models, the neural network approach tries to limit the data sparseness problem by performing the estimation in a continuous space, allowing by this means smooth interpolations. The complexity to train such a model and to calculate one n-gram probability is however several orders of magnitude higher than for the backoff models, making the new approach difficult to use in real applications. In this paper several techniques are presented that allow the use of a neural network language model in a large vocabulary speech recognition system, in particular very fast lattice rescoring and efficient training of large neural networks on training corpora of over 10 million words. The described approach achieves significant word error reductions with respect to a carefully tuned 4-gram backoff language model in a state of the art conversational speech recognizer for the DARPA rich transcriptions evaluations.
引用
收藏
页码:3059 / 3064
页数:6
相关论文
共 50 条
  • [1] Compression of recurrent neural networks for efficient language modeling
    Grachev, Artem M.
    Ignatov, Dmitry I.
    Savchenko, Andrey, V
    APPLIED SOFT COMPUTING, 2019, 79 : 354 - 362
  • [2] Efficient Communications in Training Large Scale Neural Networks
    Zhao, Yiyang
    Wang, Linnan
    Wu, Wei
    Bosilca, George
    Vuduc, Richard
    Ye, Jinmian
    Tang, Wenqi
    Xu, Zenglin
    PROCEEDINGS OF THE THEMATIC WORKSHOPS OF ACM MULTIMEDIA 2017 (THEMATIC WORKSHOPS'17), 2017, : 110 - 116
  • [3] Efficient Language Modeling with Automatic Relevance Determination in Recurrent Neural Networks
    Kodryan, Maxim
    Grachev, Artem
    Ignatov, Dmitry
    Vetrov, Dmitry
    4TH WORKSHOP ON REPRESENTATION LEARNING FOR NLP (REPL4NLP-2019), 2019, : 40 - 48
  • [4] Scalable and Efficient Training of Large Convolutional Neural Networks with Differential Privacy
    Bu, Zhiqi
    Mao, Jialin
    Xu, Shiyun
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [5] Efficient training of backpropagation neural networks
    Otair, Mohammed A.
    Salameh, Walid A.
    NEURAL NETWORK WORLD, 2006, 16 (04) : 291 - 311
  • [6] Fast and Efficient and Training of Neural Networks
    Yu, Hao
    Wilamowski
    3RD INTERNATIONAL CONFERENCE ON HUMAN SYSTEM INTERACTION, 2010, : 175 - 181
  • [7] Efficient and Reliable Training of Neural Networks
    Yu, Hao
    Wilamowski, Bogdan M.
    HSI: 2009 2ND CONFERENCE ON HUMAN SYSTEM INTERACTIONS, 2009, : 106 - 112
  • [8] LSTM Neural Networks for Language Modeling
    Sundermeyer, Martin
    Schlueter, Ralf
    Ney, Hermann
    13TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2012 (INTERSPEECH 2012), VOLS 1-3, 2012, : 194 - 197
  • [9] Neural Networks Compression for Language Modeling
    Grachev, Artem M.
    Ignatov, Dmitry I.
    Savchenko, Andrey V.
    PATTERN RECOGNITION AND MACHINE INTELLIGENCE, PREMI 2017, 2017, 10597 : 351 - 357
  • [10] Efficient Training of Large-Scale Neural Networks Using Linear Pipeline Broadcast
    University of Science and Technology, Department of Big Data Science, Daejeon
    34112, Korea, Republic of
    不详
    34141, Korea, Republic of
    不详
    34112, Korea, Republic of
    IEEE Access, 2024, (165653-165662) : 165653 - 165662