Language-model look-ahead for large vocabulary speech recognition

被引:0
|
作者
Ortmanns, S
Ney, H
Eiden, A
机构
关键词
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
In this paper, we present an efficient look-ahead technique which incorporates the language model knowledge at the earliest possible stage during the search process. This so-called language model look-ahead is built into the time synchronous beam search algorithm using a tree-organized pronunciation lexicon for a bigram language model. The language model look-ahead technique exploits the full knowledge of the bigram language model by distributing the language model probabilities over the nodes of the lexical tree for each predecessor word. We present a method for handling the resulting memory requirements. The recognition experiments performed on the 20000-word North American Business task (Nov.'96) demonstrate that in comparison with the unigram look-ahead a reduction by a factor of 5 in the acoustic search effort can be achieved without loss in recognition accuracy.
引用
收藏
页码:2095 / 2098
页数:4
相关论文
共 50 条
  • [1] Fast LM look-ahead for large vocabulary continuous speech recognition using perfect hashing
    Cardenal-López, A
    Diéguez-Tirado, FJ
    García-Mateo, C
    [J]. 2002 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOLS I-IV, PROCEEDINGS, 2002, : 705 - 708
  • [2] Efficient language model look-ahead probabilities generation using lower order LM look-ahead information
    Chen, Langzhou
    Chin, K. K.
    [J]. 2008 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, VOLS 1-12, 2008, : 4925 - 4928
  • [3] A unified language model for large vocabulary continuous speech recognition of Turkish
    Arisoy, Ebru
    Dutagaci, Helin
    Arslan, Levent M.
    [J]. SIGNAL PROCESSING, 2006, 86 (10) : 2844 - 2862
  • [4] Syllable based language model for large vocabulary continuous speech recognition of Uyghur
    [J]. Silamu, W. (wushour@xju.edu.cn), 1600, Tsinghua University (53):
  • [5] LaMemo: Language Modeling with Look-Ahead Memory
    Ji, Haozhe
    Zhang, Rongsheng
    Yang, Zhenyu
    Hu, Zhipeng
    Huang, Minlie
    [J]. NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 5747 - 5762
  • [6] Syllable Based Language Model for Large Vocabulary Continuous Speech Recognition of Polish
    Majewski, Piotr
    [J]. TEXT, SPEECH AND DIALOGUE, PROCEEDINGS, 2008, 5246 : 397 - 401
  • [7] Look-ahead memory consistency model
    Wu, CC
    Pean, DL
    Chen, C
    [J]. 1998 INTERNATIONAL CONFERENCE ON PARALLEL AND DISTRIBUTED SYSTEMS, PROCEEDINGS, 1998, : 504 - 510
  • [8] Language identification through large vocabulary continous speech recognition
    Lim, BP
    Li, HZ
    Chen, Y
    [J]. 2004 INTERNATIONAL SYMPOSIUM ON CHINESE SPOKEN LANGUAGE PROCESSING, PROCEEDINGS, 2004, : 49 - 52
  • [9] A large vocabulary continuous speech recognition system for Persian language
    Sameti, Hossein
    Veisi, Hadi
    Bahrani, Mohammad
    Babaali, Bagher
    Hosseinzadeh, Khosro
    [J]. EURASIP JOURNAL ON AUDIO SPEECH AND MUSIC PROCESSING, 2011, : 1 - 12
  • [10] Large vocabulary speech recognition with multispan statistical language models
    Bellegarda, JR
    [J]. IEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSING, 2000, 8 (01): : 76 - 84