LCMV BEAMFORMING WITH SUBSPACE PROJECTION FOR MULTI-SPEAKER SPEECH ENHANCEMENT

被引:0
|
作者
Hassani, Amin [1 ]
Bertrand, Alexander [1 ]
Moonen, Marc [1 ]
机构
[1] Katholieke Univ Leuven, Dept Elect Engn ESAT, Stadius Ctr Dynam Syst, Signal Proc & Data Analyt, Kasteelpk Arenberg 10, B-3001 Leuven, Belgium
关键词
LCMV beamforming; generalized eigen-value decomposition; subspace estimation; speech enhancement; noise reduction;
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
The linearly constrained minimum variance (LCMV) beam-former has been widely employed to extract (a mixture of) multiple desired speech signals from a collection of microphone signals, which are also polluted by other interfering speech signals and noise components. In many practical applications, the LCMV beamformer requires that the subspace corresponding to the desired and interferer signals is either known, or estimated by means of a data-driven procedure, e.g., using a generalized eigenvalue decomposition (GEVD). In practice, however, it often occurs that insufficient relevant samples are available to accurately estimate these subspaces, leading to a beamformer with poor output performance. In this paper we propose a subspace projection-based approach to improve the performance of the LCMV beamformer by exploiting the available data more efficiently. The improved performance achieved by this approach is demonstrated by means of simulation results.
引用
收藏
页码:91 / 95
页数:5
相关论文
共 50 条
  • [41] An Unsupervised Method to Select a Speaker Subset from Large Multi-Speaker Speech Synthesis Datasets
    Gallegos, Pilar Oplustil
    Williams, Jennifer
    Rownicka, Joanna
    King, Simon
    INTERSPEECH 2020, 2020, : 1758 - 1762
  • [42] Speech Rhythm-Based Speaker Embeddings Extraction from Phonemes and Phoneme Duration for Multi-Speaker Speech Synthesis
    Fujita, Kenichi
    Ando, Atsushi
    Ijima, Yusuke
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2024, E107D (01) : 93 - 104
  • [43] MULTI-SPEAKER EMOTIONAL SPEECH SYNTHESIS WITH FINE-GRAINED PROSODY MODELING
    Lu, Chunhui
    Wen, Xue
    Liu, Ruolan
    Chen, Xiao
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 5729 - 5733
  • [44] MIMO Self-attentive RNN Beamformer for Multi-speaker Speech Separation
    Li, Xiyun
    Xu, Yong
    Yu, Meng
    Zhang, Shi-Xiong
    Xu, Jiaming
    Xu, Bo
    Yu, Dong
    INTERSPEECH 2021, 2021, : 1119 - 1123
  • [45] Multi-speaker Text-to-speech Synthesis Using Deep Gaussian Processes
    Mitsui, Kentaro
    Koriyama, Tomoki
    Saruwatari, Hiroshi
    INTERSPEECH 2020, 2020, : 2032 - 2036
  • [46] J-MAC: Japanese multi-speaker audiobook corpus for speech synthesis
    Takamichi, Shinnosuke
    Nakata, Wataru
    Tanji, Naoko
    Saruwatari, Hiroshi
    Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH, 2022, 2022-September : 2358 - 2362
  • [47] J-MAC: Japanese multi-speaker audiobook corpus for speech synthesis
    Takamichi, Shinnosuke
    Nakata, Wataru
    Tanji, Naoko
    Saruwatari, Hiroshi
    INTERSPEECH 2022, 2022, : 2358 - 2362
  • [48] MULTI-SPEAKER EMOTIONAL ACOUSTIC MODELING FOR CNN-BASED SPEECH SYNTHESIS
    Choi, Heejin
    Park, Sangjun
    Park, Jinuk
    Hahn, Minsoo
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 6950 - 6954
  • [49] GANSpeech: Adversarial Training for High-Fidelity Multi-Speaker Speech Synthesis
    Yang, Jinhyeok
    Bae, Jae-Sung
    Bak, Taejun
    Kim, Young-Ik
    Cho, Hoon-Young
    INTERSPEECH 2021, 2021, : 2202 - 2206
  • [50] LIGHTSPEECH: LIGHTWEIGHT NON-AUTOREGRESSIVE MULTI-SPEAKER TEXT-TO-SPEECH
    Li, Song
    Ouyang, Beibei
    Li, Lin
    Hong, Qingyang
    2021 IEEE SPOKEN LANGUAGE TECHNOLOGY WORKSHOP (SLT), 2021, : 499 - 506