Discriminative n-gram language modeling

被引:113
|
作者
Roark, Brian
Saraclar, Murat
Collins, Michael
机构
[1] Oregon Hlth & Sci Univ, Sch Sci & Engn, Ctr Spoken Language Understanding, Beaverton, OR 97006 USA
[2] Bogazici Univ, TR-34342 Istanbul, Turkey
[3] MIT, CSAIL, EECS Stata Ctr, Cambridge, MA 02139 USA
来源
COMPUTER SPEECH AND LANGUAGE | 2007年 / 21卷 / 02期
关键词
D O I
10.1016/j.csl.2006.06.006
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper describes discriminative language modeling for a large vocabulary speech recognition task. We contrast two parameter estimation methods: the perceptron algorithm, and a method based on maximizing the regularized conditional toe-likelihood. The models are encoded as deterministic weighted finite state automata, and are applied by intersecting the automata with word-lattices that are the output from a baseline recognizer. The perceptron algorithm has the benefit of automatically selecting a relatively small feature set in just a couple of passes over the training data. We describe a method based on regularized likelihood that makes use of the feature set given by the perceptron algorithm, and initialization with the perceptron's weights; this method gives an additional 0.5% reduction in word error rate (WER) over training with the perceptron alone. The final system achieves a 1.8% absolute reduction in WER for a baseline first-pass recognition system (from 39.2% to 37.4%), and a 0.9% absolute reduction in WER for a multi-pass recognition system (from 28.9% to 28.0%). (c) 2006 Elsevier Ltd. All rights reserved.
引用
收藏
页码:373 / 392
页数:20
相关论文
共 50 条
  • [31] Comparison of phosphorylation patterns across eukaryotes by discriminative N-gram analysis
    Itziar Frades
    Svante Resjö
    Erik Andreasson
    BMC Bioinformatics, 16
  • [32] Comparison of phosphorylation patterns across eukaryotes by discriminative N-gram analysis
    Frades, Itziar
    Resjo, Svante
    Andreasson, Erik
    BMC BIOINFORMATICS, 2015, 16
  • [33] Relevance weighting for combining multi-domain data for n-gram language modeling
    Iver, R
    Ostendorf, M
    COMPUTER SPEECH AND LANGUAGE, 1999, 13 (03): : 267 - 282
  • [34] Profile based compression of n-gram language models
    Olsen, Jesper
    Oria, Daniela
    2006 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, VOLS 1-13, 2006, : 1041 - 1044
  • [35] Joint-Character-POC N-Gram Language Modeling For Chinese Speech Recognition
    Wang, Bin
    Ou, Zhijian
    Li, Jian
    Kawamura, Akinori
    2014 9TH INTERNATIONAL SYMPOSIUM ON CHINESE SPOKEN LANGUAGE PROCESSING (ISCSLP), 2014, : 24 - +
  • [36] Improved N-gram Phonotactic Models For Language Recognition
    BenZeghiba, Mohamed Faouzi
    Gauvain, Jean-Luc
    Lamel, Lori
    11TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2010 (INTERSPEECH 2010), VOLS 3 AND 4, 2010, : 2718 - 2721
  • [37] A language independent n-gram model for word segmentation
    Kang, Seung-Shik
    Hwang, Kyu-Baek
    Lect. Notes Comput. Sci., 1600, (557-565):
  • [38] A language independent n-gram model for word segmentation
    Kang, Seung-Shik
    Hwang, Kyu-Baek
    AI 2006: ADVANCES IN ARTIFICIAL INTELLIGENCE, PROCEEDINGS, 2006, 4304 : 557 - +
  • [39] N-GRAM ANALYSIS OF TEXT DOCUMENTS IN SERBIAN LANGUAGE
    Marovac, Ulfeta
    Pljaskovic, Aldina
    Crnisanin, Adela
    Kajan, Ejub
    2012 20TH TELECOMMUNICATIONS FORUM (TELFOR), 2012, : 1385 - 1388
  • [40] Efficient MDI Adaptation for n-gram Language Models
    Huang, Ruizhe
    Li, Ke
    Arora, Ashish
    Povey, Daniel
    Khudanpur, Sanjeev
    INTERSPEECH 2020, 2020, : 4916 - 4920