Self-Supervised Learning and Multi-Task Pre-Training Based Single-Channel Acoustic Denoising

被引:0
|
作者
Li, Yi [1 ]
Sun, Yang [2 ]
Naqvi, Syed Mohsen [1 ]
机构
[1] Newcastle Univ, Sch Engn, Intelligent Sensing & Commun Grp, Newcastle Upon Tyne NE1 7RU, England
[2] Univ Oxford, Big Data Inst, Oxford OX3 7LF, England
关键词
MONAURAL SOURCE SEPARATION; SPEECH; ENVIRONMENTS;
D O I
10.1109/MFI55806.2022.9913855
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In self-supervised learning-based single-channel speech denoising problem, it is challenging to reduce the gap between the denoising performance on the estimated and target speech signals with existed pre-tasks. In this paper, we propose a multi-task pre-training method to improve the speech denoising performance within self-supervised learning. In the proposed pre-training autoencoder (PAE), only a very limited set of unpaired and unseen clean speech signals are required to learn speech latent representations. Meanwhile, to solve the limitation of existing single pre-task, the proposed masking module exploits the dereverberated mask and estimated ratio mask to denoise the mixture as the new pre-task. The downstream task autoencoder (DAE) utilizes unlabeled and unseen reverberant mixtures to generate the estimated mixtures. The DAE is trained to share a latent representation with the clean examples from the learned representation in the PAE. Experimental results on a benchmark dataset demonstrate that the proposed method outperforms the state-of-the-art approaches.
引用
收藏
页数:5
相关论文
共 50 条
  • [21] A MULTI-TASK SELF-SUPERVISED LEARNING FRAMEWORK FOR SCOPY IMAGES
    Li, Yuexiang
    Chen, Jiawei
    Zheng, Yefeng
    2020 IEEE 17TH INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING (ISBI 2020), 2020, : 2005 - 2009
  • [22] Task-Customized Self-Supervised Pre-training with Scalable Dynamic Routing
    Liu, Zhili
    Han, Jianhua
    Hong, Lanqing
    Xu, Hang
    Chen, Kai
    Xu, Chunjing
    Li, Zhenguo
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 1854 - 1862
  • [23] Self-Supervised pre-training model based on Multi-view for MOOC Recommendation
    Tian R.
    Cai J.
    Li C.
    Wang J.
    Expert Systems with Applications, 2024, 252
  • [24] AN ADAPTER BASED PRE-TRAINING FOR EFFICIENT AND SCALABLE SELF-SUPERVISED SPEECH REPRESENTATION LEARNING
    Kessler, Samuel
    Thomas, Bethan
    Karout, Salah
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 3179 - 3183
  • [25] Contrast to Divide: Self-Supervised Pre-Training for Learning with Noisy Labels
    Zheltonozhskii, Evgenii
    Baskin, Chaim
    Mendelson, Avi
    Bronstein, Alex M.
    Litany, Or
    2022 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2022), 2022, : 387 - 397
  • [26] Deep learning based on self-supervised pre-training: Application on sandstone content prediction
    Wang, Chong Ming
    Wang, Xing Jian
    Chen, Yang
    Wen, Xue Mei
    Zhang, Yong Heng
    Li, Qing Wu
    FRONTIERS IN EARTH SCIENCE, 2023, 10
  • [27] Self-Supervised Pre-Training for Deep Image Prior-Based Robust PET Image Denoising
    Onishi, Yuya
    Hashimoto, Fumio
    Ote, Kibo
    Matsubara, Keisuke
    Ibaraki, Masanobu
    IEEE TRANSACTIONS ON RADIATION AND PLASMA MEDICAL SCIENCES, 2024, 8 (04) : 348 - 356
  • [28] Self-supervised Pre-training and Semi-supervised Learning for Extractive Dialog Summarization
    Zhuang, Yingying
    Song, Jiecheng
    Sadagopan, Narayanan
    Beniwal, Anurag
    COMPANION OF THE WORLD WIDE WEB CONFERENCE, WWW 2023, 2023, : 1069 - 1076
  • [29] Censer: Curriculum Semi-supervised Learning for Speech Recognition Based on Self-supervised Pre-training
    Zhang, Bowen
    Cao, Songjun
    Zhang, Xiaoming
    Zhang, Yike
    Ma, Long
    Shinozaki, Takahiro
    INTERSPEECH 2022, 2022, : 2653 - 2657
  • [30] Self-Supervised Pre-training for Time Series Classification
    Shi, Pengxiang
    Ye, Wenwen
    Qin, Zheng
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,