Efficient training of large neural networks for language modeling

被引:0
|
作者
Schwenk, H [1 ]
机构
[1] CNRS, LIMSI, F-91403 Orsay, France
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently there has been increasing interest in using neural networks for language modeling. In contrast to the well known backoff n-gram language models, the neural network approach tries to limit the data sparseness problem by performing the estimation in a continuous space, allowing by this means smooth interpolations. The complexity to train such a model and to calculate one n-gram probability is however several orders of magnitude higher than for the backoff models, making the new approach difficult to use in real applications. In this paper several techniques are presented that allow the use of a neural network language model in a large vocabulary speech recognition system, in particular very fast lattice rescoring and efficient training of large neural networks on training corpora of over 10 million words. The described approach achieves significant word error reductions with respect to a carefully tuned 4-gram backoff language model in a state of the art conversational speech recognizer for the DARPA rich transcriptions evaluations.
引用
收藏
页码:3059 / 3064
页数:6
相关论文
共 50 条
  • [21] Data-Efficient Augmentation for Training Neural Networks
    Liu, Tian Yu
    Mirzasoleiman, Baharan
    Advances in Neural Information Processing Systems, 2022, 35
  • [22] Efficient Incremental Training for Deep Convolutional Neural Networks
    Tao, Yudong
    Tu, Yuexuan
    Shyu, Mei-Ling
    2019 2ND IEEE CONFERENCE ON MULTIMEDIA INFORMATION PROCESSING AND RETRIEVAL (MIPR 2019), 2019, : 286 - 291
  • [23] An Efficient Optimization Technique for Training Deep Neural Networks
    Mehmood, Faisal
    Ahmad, Shabir
    Whangbo, Taeg Keun
    MATHEMATICS, 2023, 11 (06)
  • [24] An efficient global algorithm for supervised training of neural networks
    Shukla, KK
    Raghunath
    COMPUTERS & ELECTRICAL ENGINEERING, 1999, 25 (03) : 193 - 216
  • [25] EXODUS: Stable and efficient training of spiking neural networks
    Bauer, Felix C.
    Lenz, Gregor
    Haghighatshoar, Saeid
    Sheik, Sadique
    FRONTIERS IN NEUROSCIENCE, 2023, 17
  • [26] Efficient Training of Artificial Neural Networks for Autonomous Navigation
    Pomerleau, Dean A.
    NEURAL COMPUTATION, 1991, 3 (01) : 88 - 97
  • [27] Efficient Constructive Techniques for Training Switching Neural Networks
    Ferrari, Enrico
    Muselli, Marco
    CONSTRUCTIVE NEURAL NETWORKS, 2009, 258 : 25 - 48
  • [28] Efficient and effective training of sparse recurrent neural networks
    Liu, Shiwei
    Ni'mah, Iftitahu
    Menkovski, Vlado
    Mocanu, Decebal Constantin
    Pechenizkiy, Mykola
    NEURAL COMPUTING & APPLICATIONS, 2021, 33 (15): : 9625 - 9636
  • [29] Accurate, efficient and scalable training of Graph Neural Networks
    Zeng, Hanqing
    Zhou, Hongkuan
    Srivastava, Ajitesh
    Kannan, Rajgopal
    Prasanna, Viktor
    JOURNAL OF PARALLEL AND DISTRIBUTED COMPUTING, 2021, 147 : 166 - 183
  • [30] Data-Efficient Augmentation for Training Neural Networks
    Liu, Tian Yu
    Mirzasoleiman, Baharan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,