Discriminative n-gram language modeling

被引:113
|
作者
Roark, Brian
Saraclar, Murat
Collins, Michael
机构
[1] Oregon Hlth & Sci Univ, Sch Sci & Engn, Ctr Spoken Language Understanding, Beaverton, OR 97006 USA
[2] Bogazici Univ, TR-34342 Istanbul, Turkey
[3] MIT, CSAIL, EECS Stata Ctr, Cambridge, MA 02139 USA
来源
COMPUTER SPEECH AND LANGUAGE | 2007年 / 21卷 / 02期
关键词
D O I
10.1016/j.csl.2006.06.006
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper describes discriminative language modeling for a large vocabulary speech recognition task. We contrast two parameter estimation methods: the perceptron algorithm, and a method based on maximizing the regularized conditional toe-likelihood. The models are encoded as deterministic weighted finite state automata, and are applied by intersecting the automata with word-lattices that are the output from a baseline recognizer. The perceptron algorithm has the benefit of automatically selecting a relatively small feature set in just a couple of passes over the training data. We describe a method based on regularized likelihood that makes use of the feature set given by the perceptron algorithm, and initialization with the perceptron's weights; this method gives an additional 0.5% reduction in word error rate (WER) over training with the perceptron alone. The final system achieves a 1.8% absolute reduction in WER for a baseline first-pass recognition system (from 39.2% to 37.4%), and a 0.9% absolute reduction in WER for a multi-pass recognition system (from 28.9% to 28.0%). (c) 2006 Elsevier Ltd. All rights reserved.
引用
收藏
页码:373 / 392
页数:20
相关论文
共 50 条
  • [1] Discriminative N-gram Language Modeling for Turkish
    Arisoy, Ebru
    Roark, Brian
    Shafran, Izhak
    Saraclar, Murat
    INTERSPEECH 2008: 9TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2008, VOLS 1-5, 2008, : 825 - +
  • [2] Constrained Discriminative Training of N-gram Language Models
    Rastrow, Ariya
    Sethy, Abhinav
    Ramabhadran, Bhuvana
    2009 IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION & UNDERSTANDING (ASRU 2009), 2009, : 311 - +
  • [3] Pseudo-Conventional N-Gram Representation of the Discriminative N-Gram Model for LVCSR
    Zhou, Zhengyu
    Meng, Helen
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2010, 4 (06) : 943 - 952
  • [4] Bayesian learning of n-gram statistical language modeling
    Bai, Shuanhu
    Li, Haizhou
    2006 IEEE International Conference on Acoustics, Speech and Signal Processing, Vols 1-13, 2006, : 1045 - 1048
  • [5] Discriminative N-gram Selection for Dialect Recognition
    Richardson, F. S.
    Campbell, W. M.
    Torres-Carrasquillo, P. A.
    INTERSPEECH 2009: 10TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2009, VOLS 1-5, 2009, : 192 - 195
  • [6] Recasting the discriminative n-gram model as a pseudo-conventional n-gram model for LVCSR
    Zhou, Zhengyu
    Meng, Helen
    2008 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, VOLS 1-12, 2008, : 4933 - 4936
  • [7] NeuroYara: Learning to Rank for Yara Rules Generation Through Deep Language Modeling and Discriminative N-Gram Encoding
    Mansour, Ziad
    Ou, Weihan
    Ding, Steven H. H.
    Zulkernine, Mohammad
    Charland, Philippe
    IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, 2025, 22 (02) : 1747 - 1762
  • [8] Modeling actions of PubMed users with n-gram language models
    Lin, Jimmy
    Wilbur, W. John
    INFORMATION RETRIEVAL, 2009, 12 (04): : 487 - 503
  • [9] Modeling actions of PubMed users with n-gram language models
    Jimmy Lin
    W. John Wilbur
    Information Retrieval, 2009, 12 : 487 - 503
  • [10] Discriminative Training of n-gram Language Models for Speech Recognition via Linear Programming
    Magdin, Vladimir
    Jiang, Hui
    2009 IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION & UNDERSTANDING (ASRU 2009), 2009, : 305 - 310