Robust speech recognition method based on discriminative environment feature extraction

被引:4
|
作者
Han, JQ [1 ]
Gao, W
机构
[1] Harbin Inst Technol, Dept Comp Engn & Sci, Harbin 150001, Peoples R China
[2] Chinese Acad Sci, Inst Comp Technol, Beijing 100080, Peoples R China
基金
中国国家自然科学基金;
关键词
robust speech recognition; minimum classification error; environmental parameter; discriminative learning;
D O I
10.1007/BF02948964
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
It is an effective approach to learn the influence of environmental parameters, such as additive noise and channel distortions, from training data for robust speech recognition. Most of the previous methods are based on maximum likelihood estimation criterion. However, these methods do not lead to a minimum error rate result. In this paper, a novel discriminative learning method of environmental parameters, which is based on Minimum Classification Error (MCE) criterion, is proposed. In the method, a simple classifier and the Generalized Probabilistic Descent (GPD) algorithm are adopted to iteratively learn the environmental parameters. Consequently, the clean speech features are estimated from the noisy speech features with the estimated environmental parameters, and then the estimations of clean speech features are utilized in the back-end HMM classifier. Experiments show that the best error rate reduction of 32.1% is obtained, tested on a task of 18 isolated confusion Korean words, relative to a conventional HMM system.
引用
收藏
页码:458 / 464
页数:7
相关论文
共 50 条
  • [1] Robust speech recognition method based on discriminative environment feature extraction
    Jiqing Han
    Wen Gao
    [J]. Journal of Computer Science and Technology, 2001, 16 : 458 - 464
  • [2] Robust Speech Recognition Method Based on Discriminative Environment Feature Extraction
    韩纪庆
    高文
    [J]. Journal of Computer Science & Technology, 2001, (05) : 458 - 464
  • [3] Discriminative temporal feature extraction for robust speech recognition
    Shen, JL
    [J]. ELECTRONICS LETTERS, 1997, 33 (19) : 1598 - 1600
  • [4] Robust endpoint detection for speech recognition based on discriminative feature extraction
    Yamamoto, Koichi
    Jabloun, Firas
    Reinhard, Klaus
    Kawamura, Akinori
    [J]. 2006 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, VOLS 1-13, 2006, : 805 - 808
  • [6] A robust feature extraction based on the MTF concept for speech recognition in reverberant environment
    Lu, Xugang
    Unoki, Masashi
    Akagi, Masato
    [J]. INTERSPEECH 2006 AND 9TH INTERNATIONAL CONFERENCE ON SPOKEN LANGUAGE PROCESSING, VOLS 1-5, 2006, : 2546 - 2549
  • [7] MVDR based feature extraction for robust speech recognition
    Dharanipragada, S
    Rao, BD
    [J]. 2001 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOLS I-VI, PROCEEDINGS: VOL I: SPEECH PROCESSING 1; VOL II: SPEECH PROCESSING 2 IND TECHNOL TRACK DESIGN & IMPLEMENTATION OF SIGNAL PROCESSING SYSTEMS NEURALNETWORKS FOR SIGNAL PROCESSING; VOL III: IMAGE & MULTIDIMENSIONAL SIGNAL PROCESSING MULTIMEDIA SIGNAL PROCESSING, 2001, : 309 - 312
  • [8] A robust feature extraction method based on CZCPA model for speech recognition system
    Zhang, XY
    Jiao, ZP
    Zhao, SY
    [J]. ICEMI 2005: Conference Proceedings of the Seventh International Conference on Electronic Measurement & Instruments, Vol 3, 2005, : 89 - 92
  • [9] Feature extraction for robust speech recognition
    Dharanipragada, S
    [J]. 2002 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS, VOL II, PROCEEDINGS, 2002, : 855 - 858
  • [10] An auditory neural feature extraction method for robust speech recognition
    Guo, Wei
    Zhang, Liqing
    Xia, Bin
    [J]. 2007 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOL IV, PTS 1-3, 2007, : 793 - +