NOISE SUPPRESSION WITH UNSUPERVISED JOINT SPEAKER ADAPTATION AND NOISE MIXTURE MODEL ESTIMATION

被引:0
|
作者
Fujimoto, Masakiyo [1 ]
Watanabe, Shinji [1 ]
Nakatani, Tomohiro [1 ]
机构
[1] NTT Corp, NTT Commun Sci Labs, Seika, Kyoto 6190237, Japan
关键词
noise suppression; noise mixture model; speaker adaptation; MMSE estimation; CONTINUOUS SPEECH RECOGNITION;
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
The estimation of an accurate noise model is a crucial problem for model-based noise suppression including a vector Taylor series (VTS)-based approach. The variation of the speaker characteristics is also a crucial factor as regards the model-based noise suppression. As a result, a speaker adaptation technique plays an important role in the model-based noise suppression. To deal with former problem, we have already proposed an unsupervised estimation method for a noise mixture model. Therefore, this paper proposes a joint processing method that simultaneously achieves speaker adaptation and noise mixture model estimation. This joint processing is realized by using minimum mean squared error (MMSE) estimates of clean speech and noise. Although VTS-based approach involves non-linear transformation, the MMSE estimates make it possible to flexibly estimate accurate parameters for the joint processing without the influences of non-linear VTS transformation. In the evaluation, the proposed method provided an improvement compared with results obtained using only noise mixture model estimation.
引用
收藏
页码:4713 / 4716
页数:4
相关论文
共 50 条
  • [1] A Robust Estimation Method of Noise Mixture Model for Noise Suppression
    Fujimoto, Masakiyo
    Watanabe, Shinji
    Nakatani, Tomohiro
    12TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2011 (INTERSPEECH 2011), VOLS 1-5, 2011, : 704 - 707
  • [2] Model-based noise suppression using unsupervised estimation of hidden Markov model for non-stationary noise
    Fujimoto, Masakiyo
    Nakatani, Tomohiro
    14TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2013), VOLS 1-5, 2013, : 2981 - 2985
  • [3] Improved estimation of supervision in unsupervised speaker adaptation
    Homma, S
    Aikawa, K
    Sagayama, S
    1997 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOLS I - V: VOL I: PLENARY, EXPERT SUMMARIES, SPECIAL, AUDIO, UNDERWATER ACOUSTICS, VLSI; VOL II: SPEECH PROCESSING; VOL III: SPEECH PROCESSING, DIGITAL SIGNAL PROCESSING; VOL IV: MULTIDIMENSIONAL SIGNAL PROCESSING, NEURAL NETWORKS, 1997, : 1023 - 1026
  • [4] Unsupervised model adaptation for speaker verification
    Preti, Alexandre
    Bonastre, Jean-Francois
    INTERSPEECH 2006 AND 9TH INTERNATIONAL CONFERENCE ON SPOKEN LANGUAGE PROCESSING, VOLS 1-5, 2006, : 2090 - 2093
  • [5] Adaptation of Acoustic Models in Joint Speaker and Noise Space Using Bilinear Models
    Jeong, Yongwon
    Kim, Hyung Soon
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2014, E97D (08): : 2195 - 2199
  • [6] Joint channel estimation and phase noise suppression for OFDM systems
    Lee, JH
    Yang, JS
    Kim, SC
    Park, YW
    VTC2005-SPRING: 2005 IEEE 61ST VEHICULAR TECHNOLOGY CONFERENCE, VOLS 1-5, PROCEEDINGS, 2005, : 467 - 470
  • [7] Joint Channel Estimation and Phase Noise Suppression for OFDM Systems
    Kim, Yong-Hwa
    Kim, Seong-Cheol
    IEICE TRANSACTIONS ON COMMUNICATIONS, 2008, E91B (10) : 3371 - 3374
  • [8] Bias and noise in proportion estimation: A mixture psychophysical model
    Gouet, Camilo
    Jin, Wei
    Naiman, Daniel Q.
    Pena, Marcela
    Halberda, Justin
    COGNITION, 2021, 213
  • [9] JOINT MAP ADAPTATION OF FEATURE TRANSFORMATION AND GAUSSIAN MIXTURE MODEL FOR SPEAKER RECOGNITION
    Zhu, Donglai
    Ma, Bin
    Li, Haizhou
    2009 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOLS 1- 8, PROCEEDINGS, 2009, : 4045 - 4048
  • [10] Impact of Noise Reduction and Spectrum Estimation on Noise Robust Speaker Identification
    Godin, Keith W.
    Sadjadi, Seyed Omid
    Hansen, John H. L.
    14TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2013), VOLS 1-5, 2013, : 3623 - 3627