Discriminative n-gram language modeling

被引:113
|
作者
Roark, Brian
Saraclar, Murat
Collins, Michael
机构
[1] Oregon Hlth & Sci Univ, Sch Sci & Engn, Ctr Spoken Language Understanding, Beaverton, OR 97006 USA
[2] Bogazici Univ, TR-34342 Istanbul, Turkey
[3] MIT, CSAIL, EECS Stata Ctr, Cambridge, MA 02139 USA
来源
COMPUTER SPEECH AND LANGUAGE | 2007年 / 21卷 / 02期
关键词
D O I
10.1016/j.csl.2006.06.006
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper describes discriminative language modeling for a large vocabulary speech recognition task. We contrast two parameter estimation methods: the perceptron algorithm, and a method based on maximizing the regularized conditional toe-likelihood. The models are encoded as deterministic weighted finite state automata, and are applied by intersecting the automata with word-lattices that are the output from a baseline recognizer. The perceptron algorithm has the benefit of automatically selecting a relatively small feature set in just a couple of passes over the training data. We describe a method based on regularized likelihood that makes use of the feature set given by the perceptron algorithm, and initialization with the perceptron's weights; this method gives an additional 0.5% reduction in word error rate (WER) over training with the perceptron alone. The final system achieves a 1.8% absolute reduction in WER for a baseline first-pass recognition system (from 39.2% to 37.4%), and a 0.9% absolute reduction in WER for a multi-pass recognition system (from 28.9% to 28.0%). (c) 2006 Elsevier Ltd. All rights reserved.
引用
收藏
页码:373 / 392
页数:20
相关论文
共 50 条
  • [21] A New Estimate of the n-gram Language Model
    Aouragh, Si Lhoussain
    Yousfi, Abdellah
    Laaroussi, Saida
    Gueddah, Hicham
    Nejja, Mohammed
    AI IN COMPUTATIONAL LINGUISTICS, 2021, 189 : 211 - 215
  • [22] Perplexity of n-Gram and Dependency Language Models
    Popel, Martin
    Marecek, David
    TEXT, SPEECH AND DIALOGUE, 2010, 6231 : 173 - 180
  • [23] MIXTURE OF MIXTURE N-GRAM LANGUAGE MODELS
    Sak, Hasim
    Allauzen, Cyril
    Nakajima, Kaisuke
    Beaufays, Francoise
    2013 IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING (ASRU), 2013, : 31 - 36
  • [24] A variant of n-gram based language classification
    Tomovic, Andrija
    Janicic, Predrag
    AI(ASTERISK)IA 2007: ARTIFICIAL INTELLIGENCE AND HUMAN-ORIENTED COMPUTING, 2007, 4733 : 410 - +
  • [25] Development of the N-gram Model for Azerbaijani Language
    Bannayeva, Aliya
    Aslanov, Mustafa
    2020 IEEE 14TH INTERNATIONAL CONFERENCE ON APPLICATION OF INFORMATION AND COMMUNICATION TECHNOLOGIES (AICT2020), 2020,
  • [26] MLP emulation of N-gram models as a first step to connectionist language modeling
    Castro, MJ
    Prat, F
    Casacuberta, F
    NINTH INTERNATIONAL CONFERENCE ON ARTIFICIAL NEURAL NETWORKS (ICANN99), VOLS 1 AND 2, 1999, (470): : 910 - 915
  • [27] Improving N-gram Language Modeling for Code-switching Speech Recognition
    Zeng, Zhiping
    Xu, Haihua
    Chong, Tze Yuang
    Chng, Eng-Siong
    Li, Haizhou
    2017 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC 2017), 2017, : 1546 - 1551
  • [28] ERNIE-Gram: Pre-Training with Explicitly N-Gram Masked Language Modeling for Natural Language Understanding
    Xiao, Dongling
    Li, Yu-Kun
    Zhang, Han
    Sun, Yu
    Tian, Hao
    Wu, Hua
    Wang, Haifeng
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 1702 - 1715
  • [29] A discriminative method for protein remote homology detection based on N-Gram
    Xie, S.
    Li, P.
    Jiang, Y.
    Zhao, Y.
    GENETICS AND MOLECULAR RESEARCH, 2015, 14 (01): : 69 - 78
  • [30] n-Gram Geo-trace Modeling
    Buthpitiya, Senaka
    Zhang, Ying
    Dey, Anind K.
    Griss, Martin
    PERVASIVE COMPUTING, 2011, 6696 : 97 - 114