Source-Filter-Based Single-Channel Speech Separation Using Pitch Information

被引:38
|
作者
Stark, Michael [1 ]
Wohlmayr, Michael [1 ]
Pernkopf, Franz [1 ]
机构
[1] Graz Univ Technol, Signal Proc & Speech Commun Lab, A-8010 Graz, Austria
基金
奥地利科学基金会;
关键词
Single-channel speech separation (SCSS); multi-pitch estimation; source-filter representation; ALGORITHM; TRACKING;
D O I
10.1109/TASL.2010.2047419
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
In this paper, we investigate the source-filter-based approach for single-channel speech separation. We incorporate source-driven aspects by multi-pitch estimation in the model-driven method. For multi-pitch estimation, the factorial HMM is utilized. For modeling the vocal tract filters either vector quantization (VQ) or non-negative matrix factorization are considered. For both methods, the final combination of the source and filter model results in an utterance dependent model that finally enables speaker independent source separation. The contributions of the paper are the multi-pitch tracker, the gain estimation for the VQ based method which accounts for different mixing levels, and a fast approximation for the likelihood computation. Additionally, a linear relationship between pitch tracking performance and speech separation performance is shown.
引用
下载
收藏
页码:242 / 255
页数:14
相关论文
共 50 条
  • [31] Speaker Verification-Based Evaluation of Single-Channel Speech Separation
    Maciejewski, Matthew
    Watanabe, Shinji
    Khudanpur, Sanjeev
    INTERSPEECH 2021, 2021, : 3520 - 3524
  • [32] Single-channel Multi-speakers Speech Separation Based on Isolated Speech Segments
    Shanfa Ke
    Zhongyuan Wang
    Ruimin Hu
    Xiaochen Wang
    Neural Processing Letters, 2023, 55 : 385 - 400
  • [33] Single-channel Multi-speakers Speech Separation Based on Isolated Speech Segments
    Ke, Shanfa
    Wang, Zhongyuan
    Hu, Ruimin
    Wang, Xiaochen
    NEURAL PROCESSING LETTERS, 2023, 55 (01) : 385 - 400
  • [34] A high SIR Low-overhead Implementation of Single-channel Speech Source Separation
    Nwaogo, Lawrence
    Bjorkqvist, Jerker
    2022 IEEE 12TH SENSOR ARRAY AND MULTICHANNEL SIGNAL PROCESSING WORKSHOP (SAM), 2022, : 440 - 444
  • [35] JOINT SINGLE-CHANNEL SPEECH SEPARATION AND SPEAKER IDENTIFICATION
    Mowlaee, P.
    Saeidi, R.
    Tan, Z. -H.
    Christensen, M. G.
    Franti, P.
    Jensen, S. H.
    2010 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2010, : 4430 - 4433
  • [36] WHAMR!: NOISY AND REVERBERANT SINGLE-CHANNEL SPEECH SEPARATION
    Maciejewski, Matthew
    Wichern, Gordon
    McQuinn, Emmett
    Le Roux, Jonathan
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 696 - 700
  • [37] LEARNING A HIERARCHICAL DICTIONARY FOR SINGLE-CHANNEL SPEECH SEPARATION
    Bao, Guangzhao
    Xu, Yangfei
    Xu, Xu
    Ye, Zhongfu
    2014 IEEE WORKSHOP ON STATISTICAL SIGNAL PROCESSING (SSP), 2014, : 476 - 479
  • [38] Learning a Discriminative Dictionary for Single-Channel Speech Separation
    Bao, Guangzhao
    Xu, Yangfei
    Ye, Zhongfu
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2014, 22 (07) : 1130 - 1138
  • [39] Improved Phase Reconstruction in Single-Channel Speech Separation
    Mayer, Florian
    Mowlaee, Pejman
    16TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2015), VOLS 1-5, 2015, : 1795 - 1799
  • [40] Single-Channel Speech Separation Focusing on Attention DE
    Li, Xinshu
    Tan, Zhenhua
    Xia, Zhenche
    Wu, Danke
    Zhang, Bin
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 3204 - 3209