SKILL: SIMILARITY-AWARE KNOWLEDGE DISTILLATION FOR SPEECH SELF-SUPERVISED LEARNING

被引:0
|
作者
Zampierin, Luca [1 ,2 ]
Hacene, Ghouthi Boukli [1 ,5 ]
Nguyen, Bac [1 ]
Ravanelli, Mirco [3 ,4 ,5 ]
机构
[1] Sony Europe BV, Stuttgart Lab 1, Stuttgart, Germany
[2] Ecole Polytech Fed Lausanne, Lausanne, Switzerland
[3] Concordia Univ, Montreal, PQ, Canada
[4] Univ Montreal, Montreal, PQ, Canada
[5] Mila Quebec AI Inst, Montreal, PQ, Canada
关键词
Model compression; self-supervised learning; knowledge distillation;
D O I
10.1109/ICASSPW62465.2024.10626978
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Self-supervised learning (SSL) has achieved remarkable success across various speech-processing tasks. To enhance its efficiency, previous works often leverage the use of compression techniques. A notable recent attempt is DPHuBERT, which applies joint knowledge distillation (KD) and structured pruning to learn a significantly smaller SSL model. In this paper, we contribute to this research domain by introducing SKILL, a novel method that conducts distillation across groups of layers instead of distilling individual arbitrarily selected layers within the teacher network. The identification of the layers to distill is achieved through a hierarchical clustering procedure applied to layer similarity measures. Extensive experiments demonstrate that our distilled version ofWavLM Base+ not only outperforms DPHuBERT but also achieves state-of-the-art results in the 30M parameters model class across several SUPERB tasks.
引用
收藏
页码:675 / 679
页数:5
相关论文
共 50 条
  • [31] Multi-Mode Online Knowledge Distillation for Self-Supervised Visual Representation Learning
    Song, Kaiyou
    Xie, Jin
    Zhang, Shan
    Luo, Zimeng
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 11848 - 11857
  • [32] Remote Sensing Image Scene Classification via Self-Supervised Learning and Knowledge Distillation
    Zhao, Yibo
    Liu, Jianjun
    Yang, Jinlong
    Wu, Zebin
    REMOTE SENSING, 2022, 14 (19)
  • [33] Improving Self-supervised Lightweight Model Learning via Hard-Aware Metric Distillation
    Liu, Hao
    Ye, Mang
    COMPUTER VISION, ECCV 2022, PT XXXI, 2022, 13691 : 295 - 311
  • [34] Self-supervised Knowledge Distillation Using Singular Value Decomposition
    Lee, Seung Hyun
    Kim, Dae Ha
    Song, Byung Cheol
    COMPUTER VISION - ECCV 2018, PT VI, 2018, 11210 : 339 - 354
  • [35] SSSD: Self-Supervised Self Distillation
    Chen, Wei-Chi
    Chu, Wei-Ta
    2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, : 2769 - 2776
  • [36] Self-Supervised Reinforcement Learning with dual-reward for knowledge-aware recommendation
    Zhang, Wei
    Lin, Yuanguo
    Liu, Yong
    You, Huanyu
    Wu, Pengcheng
    Lin, Fan
    Zhou, Xiuze
    APPLIED SOFT COMPUTING, 2022, 131
  • [37] Knowledge-aware reasoning with self-supervised reinforcement learning for explainable recommendation in MOOCs
    Lin, Yuanguo
    Zhang, Wei
    Lin, Fan
    Zeng, Wenhua
    Zhou, Xiuze
    Wu, Pengcheng
    NEURAL COMPUTING & APPLICATIONS, 2024, 36 (08): : 4115 - 4132
  • [38] Knowledge-aware reasoning with self-supervised reinforcement learning for explainable recommendation in MOOCs
    Yuanguo Lin
    Wei Zhang
    Fan Lin
    Wenhua Zeng
    Xiuze Zhou
    Pengcheng Wu
    Neural Computing and Applications, 2024, 36 : 4115 - 4132
  • [39] Structure-aware protein self-supervised learning
    Chen, Can
    Zhou, Jingbo
    Wang, Fan
    Liu, Xue
    Dou, Dejing
    BIOINFORMATICS, 2023, 39 (04)
  • [40] Self-Supervised Attention-Aware Reinforcement Learning
    Wu, Haiping
    Khetarpa, Khimya
    Precup, Doina
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 10311 - 10319