A two-stage method for single-channel speech enhancement

被引:2
|
作者
Hamid, ME [1 ]
Fukabayashi, T
机构
[1] Shizuoka Univ, Grad Sch Elect Sci & Technol, Hamamatsu, Shizuoka 4328561, Japan
[2] Shizuoka Univ, Fac Engn, Hamamatsu, Shizuoka 4328561, Japan
关键词
enhancement of speech; single-channel; autocorrelalion function; degree of noise; subtraction in time domain; blind source separation;
D O I
10.1093/ietfec/e89-a.4.1058
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
A time domain (TD) speech enhancement technique to improve SNR in noise-contaminated speech is proposed. Additional supplementary scheme is applied to estimate the degree of noise of noisy speech. This is estimated from a function, which is previously prepared as the function of the parameter of the degree of noise. The function is obtained by least square (LS) method using the given degree of noise and the estimated parameter of the degree of noise. This parameter is obtained from the autocorrelation function (ACF) on frame-by-frame basis. This estimator almost accurately estimates the degree of noise and it is useful to reduce noise. The proposed method is based on two-stage processing. In the first stage, subtraction in time domain (STD), which is equivalent to ordinary spectral subtraction (SS), is carried out. In the result, the noise is reduced to a certain level. Further reduction of noise and by-product noise residual is carried out in the second stage, where blind source separation (BSS) technique is applied in time domain. Because the method is a single-channel speech enhancement, the other signal is generated by taking the noise characteristics into consideration in order to apply BSS. The generated signal plays a very important role in BSS. This paper presents an adaptive algorithm for separating sources in convolutive mixtures modeled by finite impulse response (FIR) filters. The coefficients of the FIR filter are estimated from the decorrelation of two mixtures. Here we are recovering only one signal of interest, in particular the voice of primary speaker free from interfering noises. In the experiment, the different levels of noise are added to the clean speech signal and the improvement of SNR at each stage is investigated. The noise types considered initially in this study consist of the synthesized white and color noise with SNR set from 0 to 30 dB. The proposed method is also tested with other real-world noises. The results show that the satisfactory SNR improvement is attained in the two-stage processing.
引用
收藏
页码:1058 / 1068
页数:11
相关论文
共 50 条
  • [21] A spectral conversion approach to single-channel speech enhancement
    Mouchtaris, Athanasios
    Van der Spiegel, Jan
    Mueller, Paul
    Tsakalides, Panagiotis
    IEEE TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2007, 15 (04): : 1180 - 1193
  • [22] A two-stage algorithm for enhancement of reverberant speech
    Wu, MY
    Wang, D
    2005 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOLS 1-5: SPEECH PROCESSING, 2005, : 1085 - 1088
  • [23] TWO-STAGE DATA-DRIVEN SINGLE CHANNEL SPEECH ENHANCEMENT WITH CEPSTRAL ANALYSIS PRE-PROCESSING
    Rao, Yu
    Vahanesa, Chetan
    Reddy, Chandan K. A.
    Panahi, Issa M. S.
    2015 IEEE GLOBAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (GLOBALSIP), 2015, : 702 - 706
  • [24] A robust two-stage sleep spindle detection approach using single-channel EEG
    Jiang, Dihong
    Ma, Yu
    Wang, Yuanyuan
    JOURNAL OF NEURAL ENGINEERING, 2021, 18 (02)
  • [25] Two-Stage Single-Channel Audio Source Separation Using Deep Neural Networks
    Grais, Emad M.
    Roma, Gerard
    Simpson, Andrew J. R.
    Plumbley, Mark D.
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2017, 25 (09) : 1469 - 1479
  • [26] Single-channel Speech Enhancement Using Graph Fourier Transform
    Zhang, Chenhui
    Pan, Xiang
    INTERSPEECH 2022, 2022, : 946 - 950
  • [27] Hybrid quality measures for single-channel speech enhancement algorithms
    Dreiseitel, P
    EUROPEAN TRANSACTIONS ON TELECOMMUNICATIONS, 2002, 13 (02): : 159 - 165
  • [28] Single-channel multiple regression for in-car speech enhancement
    Li, WF
    Itou, K
    Takeda, K
    Itakura, F
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2006, E89D (03) : 1032 - 1039
  • [29] Combine Waveform and Spectral Methods for Single-channel Speech Enhancement
    Li, Miao
    Zhang, Hui
    Zhang, Xueliang
    PROCEEDINGS OF 2022 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2022, : 47 - 52
  • [30] Single-channel speech enhancement based on frequency domain ALE
    Nakanishi, Isao
    Nagata, Yuudai
    Itoh, Yoshio
    Fukui, Yutaka
    2006 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS, VOLS 1-11, PROCEEDINGS, 2006, : 2541 - 2544