Discriminative n-gram language modeling

被引:113
|
作者
Roark, Brian
Saraclar, Murat
Collins, Michael
机构
[1] Oregon Hlth & Sci Univ, Sch Sci & Engn, Ctr Spoken Language Understanding, Beaverton, OR 97006 USA
[2] Bogazici Univ, TR-34342 Istanbul, Turkey
[3] MIT, CSAIL, EECS Stata Ctr, Cambridge, MA 02139 USA
来源
COMPUTER SPEECH AND LANGUAGE | 2007年 / 21卷 / 02期
关键词
D O I
10.1016/j.csl.2006.06.006
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper describes discriminative language modeling for a large vocabulary speech recognition task. We contrast two parameter estimation methods: the perceptron algorithm, and a method based on maximizing the regularized conditional toe-likelihood. The models are encoded as deterministic weighted finite state automata, and are applied by intersecting the automata with word-lattices that are the output from a baseline recognizer. The perceptron algorithm has the benefit of automatically selecting a relatively small feature set in just a couple of passes over the training data. We describe a method based on regularized likelihood that makes use of the feature set given by the perceptron algorithm, and initialization with the perceptron's weights; this method gives an additional 0.5% reduction in word error rate (WER) over training with the perceptron alone. The final system achieves a 1.8% absolute reduction in WER for a baseline first-pass recognition system (from 39.2% to 37.4%), and a 0.9% absolute reduction in WER for a multi-pass recognition system (from 28.9% to 28.0%). (c) 2006 Elsevier Ltd. All rights reserved.
引用
收藏
页码:373 / 392
页数:20
相关论文
共 50 条
  • [41] N-gram language models for massively parallel devices
    Bogoychev, Nikolay
    Lopez, Adam
    PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, 2016, : 1944 - 1953
  • [42] Language Identification based on n-gram Frequency Ranking
    Cordoba, R.
    D'Haro, L. F.
    Fernandez-Martinez, F.
    Macias-Guarasa, J.
    Ferreiros, J.
    INTERSPEECH 2007: 8TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION, VOLS 1-4, 2007, : 1921 - 1924
  • [43] Multilingual stochastic n-gram class language models
    Jardino, M
    1996 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, CONFERENCE PROCEEDINGS, VOLS 1-6, 1996, : 161 - 163
  • [44] POWER LAW DISCOUNTING FOR N-GRAM LANGUAGE MODELS
    Huang, Songfang
    Renals, Steve
    2010 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2010, : 5178 - 5181
  • [45] N-gram language models for document image decoding
    Kopec, GE
    Said, MR
    Popat, K
    DOCUMENT RECOGNITION AND RETRIEVAL IX, 2002, 4670 : 191 - 202
  • [46] Bugram: Bug Detection with N-gram Language Models
    Wang, Song
    Chollak, Devin
    Movshovitz-Attias, Dana
    Tan, Lin
    2016 31ST IEEE/ACM INTERNATIONAL CONFERENCE ON AUTOMATED SOFTWARE ENGINEERING (ASE), 2016, : 708 - 719
  • [47] Modified Chinese N-gram statistical language model
    Tian, Bin
    Tian, Hongxin
    Yi, Kechu
    Xi'an Dianzi Keji Daxue Xuebao/Journal of Xidian University, 2000, 27 (01): : 62 - 64
  • [48] Active Learning for Language Identification with N-gram Technique
    Feng , Yuxin
    2021 2ND INTERNATIONAL CONFERENCE ON BIG DATA & ARTIFICIAL INTELLIGENCE & SOFTWARE ENGINEERING (ICBASE 2021), 2021, : 560 - 564
  • [49] SUITE OF TOOLS FOR STATISTICAL N-GRAM LANGUAGE MODELING FOR PATTERN MINING IN WHOLE GENOME SEQUENCES
    Ganapathiraju, Madhavi K.
    Mitchell, Asia D.
    Thahir, Mohamed
    Motwani, Kamiya
    Ananthasubramanian, Seshan
    JOURNAL OF BIOINFORMATICS AND COMPUTATIONAL BIOLOGY, 2012, 10 (06)
  • [50] N-gram Insight
    Prans, George
    AMERICAN SCIENTIST, 2011, 99 (05) : 356 - 357