Efficient training of large neural networks for language modeling

被引:0
|
作者
Schwenk, H [1 ]
机构
[1] CNRS, LIMSI, F-91403 Orsay, France
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently there has been increasing interest in using neural networks for language modeling. In contrast to the well known backoff n-gram language models, the neural network approach tries to limit the data sparseness problem by performing the estimation in a continuous space, allowing by this means smooth interpolations. The complexity to train such a model and to calculate one n-gram probability is however several orders of magnitude higher than for the backoff models, making the new approach difficult to use in real applications. In this paper several techniques are presented that allow the use of a neural network language model in a large vocabulary speech recognition system, in particular very fast lattice rescoring and efficient training of large neural networks on training corpora of over 10 million words. The described approach achieves significant word error reductions with respect to a carefully tuned 4-gram backoff language model in a state of the art conversational speech recognizer for the DARPA rich transcriptions evaluations.
引用
收藏
页码:3059 / 3064
页数:6
相关论文
共 50 条
  • [11] Efficient training of unitary optical neural networks
    Lu, Kunrun
    Guo, Xianxin
    OPTICS EXPRESS, 2023, 31 (24) : 39616 - 39623
  • [12] Efficient training for dendrite morphological neural networks
    Sossa, Humberto
    Guevara, Elizabeth
    NEUROCOMPUTING, 2014, 131 : 132 - 142
  • [13] TDTO Language Modeling with Feedforward Neural Networks
    Chong, Tze Yuang
    Banchs, Rafael E.
    Chng, Eng Siong
    Li, Haizhou
    16TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2015), VOLS 1-5, 2015, : 1458 - 1462
  • [14] SLONN - A SIMULATION LANGUAGE FOR MODELING OF NEURAL NETWORKS
    WANG, DL
    HSU, C
    SIMULATION, 1990, 55 (02) : 69 - 83
  • [15] Sequential recurrent neural networks for language modeling
    Oualil, Youssef
    Greenberg, Clayton
    Singh, Mittul
    Klakow, Dietrich
    17TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2016), VOLS 1-5: UNDERSTANDING SPEECH PROCESSING IN HUMANS AND MACHINES, 2016, : 3509 - 3513
  • [16] Efficient training of interval Neural Networks for imprecise training data
    Sadeghi, Jonathan
    de Angelis, Marco
    Patelli, Edoardo
    NEURAL NETWORKS, 2019, 118 : 338 - 351
  • [17] Strategies for Training Large Vocabulary Neural Language Models
    Chen, Wenlin
    Grangier, David
    Auli, Michael
    PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, 2016, : 1975 - 1985
  • [18] Improving Neural Language Modeling via Adversarial Training
    Wang, Dilin
    Gong, Chengyue
    Liu, Qiang
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [19] Recurrent Neural Networks and Morphological Features in Language Modeling Features in Language Modeling for Serbian
    Pakoci, Edvin T.
    Popovic, Branislav Z.
    2021 29TH TELECOMMUNICATIONS FORUM (TELFOR), 2021,
  • [20] Efficient and effective training of sparse recurrent neural networks
    Shiwei Liu
    Iftitahu Ni’mah
    Vlado Menkovski
    Decebal Constantin Mocanu
    Mykola Pechenizkiy
    Neural Computing and Applications, 2021, 33 : 9625 - 9636