An effective cluster-based model for robust speech detection and speech recognition in noisy environments

被引:21
|
作者
Gorriz, J. M. [1 ]
Ramirez, J.
Segura, J. C.
Puntonet, C. G.
机构
[1] Univ Granada, Dept Signal Theory, Granada, Spain
[2] Univ Granada, Dept Comp Architecture & Technol, Granada, Spain
来源
关键词
D O I
10.1121/1.2208450
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
This paper shows an accurate speech detection algorithm for improving the performance of speech recognition systems working in noisy environments. The proposed method is based on a hard decision clustering approach where a set of prototypes is used to characterize the noisy channel. Detecting the presence of speech is enabled by a decision rule formulated in terms of an averaged distance between the observation vector and a cluster-based noise model. The algorithm benefits from using contextual information, a strategy that considers not only a single speech frame but also a neighborhood of data in order to smooth the decision function and improve speech detection robustness. The proposed scheme exhibits reduced computational cost making it adequate for real time applications, i.e., automated speech recognition systems. An exhaustive analysis is conducted on the AURORA 2 and AURORA 3 databases in order to assess the performance of the algorithm and to compare it to existing standard voice activity detection (VAD) methods. The results show significant improvements in detection accuracy and speech recognition rate over standard VADs such as ITU-T G.729, ETSI GSM AMR, and ETSI AFE for distributed speech recognition and a representative set of recently reported VAD algorithms. (c) 2006 Acoustical Society of America.
引用
收藏
页码:470 / 481
页数:12
相关论文
共 50 条
  • [31] Speech Emotion Recognition in Noisy and Reverberant Environments
    Heracleous, Panikos
    Yasuda, Keiji
    Sugaya, Fumiaki
    Yoneyama, Akio
    Hashimoto, Masayuki
    2017 SEVENTH INTERNATIONAL CONFERENCE ON AFFECTIVE COMPUTING AND INTELLIGENT INTERACTION (ACII), 2017, : 262 - 266
  • [32] Multisensory benefits for speech recognition in noisy environments
    Oh, Yonghee
    Schwalm, Meg
    Kalpin, Nicole
    FRONTIERS IN NEUROSCIENCE, 2022, 16
  • [33] Speech Recognition On Mobile Devices In Noisy Environments
    Yurtcan, Yaser
    Kilic, Banu Gunel
    2018 26TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2018,
  • [34] Robust speech recognition in noisy environments: The 2001 IBM SPINE evaluation system
    Kingsbury, B
    Saon, G
    Mangu, L
    Padmanabhan, M
    Sarikaya, R
    2002 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOLS I-IV, PROCEEDINGS, 2002, : 53 - 56
  • [35] Robust Arabic speech recognition in noisy environments using prosodic features and formant
    Amrous, Anissa
    Debyeche, Mohamed
    Amrouche, Abderrahman
    INTERNATIONAL JOURNAL OF SPEECH TECHNOLOGY, 2011, 14 (04) : 351 - 359
  • [36] Speech enhancement strategy for speech recognition microcontroller under noisy environments
    Chan, Kit Yan
    Nordholm, Sven
    Yiu, Ka Fai Cedric
    Togneri, Roberto
    NEUROCOMPUTING, 2013, 118 : 279 - 288
  • [37] Robust speech recognition in car environments
    Shozakai, M
    Nakamura, S
    Shikano, K
    PROCEEDINGS OF THE 1998 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, VOLS 1-6, 1998, : 269 - 272
  • [38] Robust speech endpoint detection based on MP3 file in various noisy environments
    Wang, Fang
    Huang, Xianglin
    Yang, Lifang
    Liu, Tao
    2008 INTERNATIONAL CONFERENCE ON AUDIO, LANGUAGE AND IMAGE PROCESSING, VOLS 1 AND 2, PROCEEDINGS, 2008, : 670 - 675
  • [39] Face-to-talk: Audio-visual speech detection for robust speech recognition in noisy environment
    Murai, K
    Nakamura, S
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2003, E86D (03): : 505 - 513
  • [40] Advancing Speech Recognition With No Speech Or With Noisy Speech
    Krishna, Gautam
    Tran, Co
    Carnahan, Mason
    Tewfik, Ahmed
    2019 27TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO), 2019,