Compression of Acoustic Model via Knowledge Distillation and Pruning

被引:0
|
作者
Li, Chenxing [1 ,2 ]
Zhu, Lei [3 ]
Xu, Shuang [1 ]
Gao, Peng [3 ]
Xu, Bo [1 ]
机构
[1] Chinese Acad Sci, Inst Automat, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Beijing, Peoples R China
[3] Rokid Inc Beijing, AI Lab, Beijing, Peoples R China
关键词
speech recognition; layer normalization; knowledge distillation; pruning; model compression;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, the performance of speech recognition system based on neural network has been greatly improved. Arguably, this huge improvement can be mainly attributed to deeper and wider layers. These systems are more difficult to be deployed on the embedded devices due to their large size and high computational complexity. To address these issues, we propose a method to compress deep feed-forward neural network (DNN) based acoustic model. In detail, a state-of-the-art acoustic model is trained as the baseline model. In this step, layer normalization is applied to accelerating the model convergence and improving the generalization performance. Knowledge distillation and pruning are then conducted to compress the model. Our final model can achieve 14:59x parameters reduction, 5x storage size reduction and comparable performance compared with the baseline model.
引用
收藏
页码:2785 / 2790
页数:6
相关论文
共 50 条
  • [1] PQK: Model Compression via Pruning, Quantization, and Knowledge Distillation
    Kim, Jangho
    Chang, Simyung
    Kwak, Nojun
    [J]. INTERSPEECH 2021, 2021, : 4568 - 4572
  • [2] Model compression via pruning and knowledge distillation for person re-identification
    Xie, Haonan
    Jiang, Wei
    Luo, Hao
    Yu, Hongyan
    [J]. JOURNAL OF AMBIENT INTELLIGENCE AND HUMANIZED COMPUTING, 2021, 12 (02) : 2149 - 2161
  • [3] Model compression via pruning and knowledge distillation for person re-identification
    Haonan Xie
    Wei Jiang
    Hao Luo
    Hongyan Yu
    [J]. Journal of Ambient Intelligence and Humanized Computing, 2021, 12 : 2149 - 2161
  • [4] Private Model Compression via Knowledge Distillation
    Wang, Ji
    Bao, Weidong
    Sun, Lichao
    Zhu, Xiaomin
    Cao, Bokai
    Yu, Philip S.
    [J]. THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 1190 - +
  • [5] Efficient and Controllable Model Compression through Sequential Knowledge Distillation and Pruning
    Malihi, Leila
    Heidemann, Gunther
    [J]. BIG DATA AND COGNITIVE COMPUTING, 2023, 7 (03)
  • [6] Model Compression by Iterative Pruning with Knowledge Distillation and Its Application to Speech Enhancement
    Wei, Zeyuan
    Li, Hao
    Zhang, Xueliang
    [J]. INTERSPEECH 2022, 2022, : 941 - 945
  • [7] Joint structured pruning and dense knowledge distillation for efficient transformer model compression
    Cui, Baiyun
    Li, Yingming
    Zhang, Zhongfei
    [J]. NEUROCOMPUTING, 2021, 458 : 56 - 69
  • [8] Combining Weight Pruning and Knowledge Distillation For CNN Compression
    Aghli, Nima
    Ribeiro, Eraldo
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2021, 2021, : 3185 - 3192
  • [9] Model Compression Algorithm via Reinforcement Learning and Knowledge Distillation
    Liu, Botao
    Hu, Bing-Bing
    Zhao, Ming
    Peng, Sheng-Lung
    Chang, Jou-Ming
    Tsoulos, Ioannis G.
    [J]. MATHEMATICS, 2023, 11 (22)
  • [10] Matching the Ideal Pruning Method with Knowledge Distillation for Optimal Compression
    Malihi, Leila
    Heidemann, Gunther
    [J]. APPLIED SYSTEM INNOVATION, 2024, 7 (04)