Stage-Wise Magnitude-Based Pruning for Recurrent Neural Networks

被引:0
|
作者
Li, Guiying [1 ,2 ]
Yang, Peng [2 ,3 ]
Qian, Chao [4 ]
Hong, Richang [5 ]
Tang, Ke [1 ,2 ]
机构
[1] Southern Univ Sci & Technol, Res Inst Trustworthy Autonomous Syst, Shenzhen 518055, Peoples R China
[2] Southern Univ Sci & Technol, Dept Comp Sci & Engn, Guangdong Key Lab Brain Inspired Intelligent Comp, Shenzhen 518055, Peoples R China
[3] Southern Univ Sci & Technol, Dept Stat & Data Sci, Shenzhen 518055, Peoples R China
[4] Nanjing Univ, State Key Lab Novel Software Technol, Nanjing 210023, Peoples R China
[5] Hefei Univ Technol, Sch Comp & Informat, Hefei 230009, Peoples R China
基金
中国国家自然科学基金;
关键词
Computational modeling; Recurrent neural networks; Neurons; Sparse matrices; Task analysis; Hardware; Training; Deep neural network (DNN) compression; language models; machine translation; optimization; recurrent neural networks (RNNs); ACCELERATION; COMPRESSION;
D O I
10.1109/TNNLS.2022.3184730
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A recurrent neural network (RNN) has shown powerful performance in tackling various natural language processing (NLP) tasks, resulting in numerous powerful models containing both RNN neurons and feedforward neurons. On the other hand, the deep structure of RNN has heavily restricted its implementation on mobile devices, where quite a few applications involve NLP tasks. Magnitude-based pruning (MP) is a promising way to address such a challenge. However, the existing MP methods are mostly designed for feedforward neural networks that do not involve a recurrent structure, and, thus, have performed less satisfactorily on pruning models containing RNN layers. In this article, a novel stage-wise MP method is proposed by explicitly taking the featured recurrent structure of RNN into account, which can effectively prune feedforward layers and RNN layers, simultaneously. The connections of neural networks are first grouped into three types according to how they are intersected with recurrent neurons. Then, an optimization-based pruning method is applied to compress each group of connections, respectively. Empirical studies show that the proposed method performs significantly better than the commonly used RNN pruning methods; i.e., up to 96.84% connections are pruned with little or even no degradation of precision indicators on the testing datasets.
引用
收藏
页码:1666 / 1680
页数:15
相关论文
共 50 条
  • [1] Optimization based Layer-wise Magnitude-based Pruning for DNN Compression
    Li, Guiying
    Qian, Chao
    Jiang, Chunhui
    Lu, Xiaofen
    Tang, Ke
    [J]. PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2018, : 2383 - 2389
  • [2] Effective Model Compression via Stage-wise Pruning
    Zhang, Ming-Yang
    Yu, Xin-Yi
    Ou, Lin-Lin
    [J]. MACHINE INTELLIGENCE RESEARCH, 2023, 20 (06) : 937 - 951
  • [3] Effective Model Compression via Stage-wise Pruning
    Ming-Yang Zhang
    Xin-Yi Yu
    Lin-Lin Ou
    [J]. Machine Intelligence Research, 2023, 20 : 937 - 951
  • [4] Stage-Wise Neural Architecture Search
    Jordao, Artur
    Akio, Fernando
    Lie, Maiko
    Schwartz, William Robson
    [J]. 2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 1985 - 1992
  • [5] LAYER-ADAPTIVE SPARSITY FOR THE MAGNITUDE-BASED PRUNING
    Lee, Jaeho
    Park, Sejun
    Mo, Sangwoo
    Ahn, Sungsoo
    Shin, Jinwoo
    [J]. ICLR 2021 - 9th International Conference on Learning Representations, 2021,
  • [6] Gradient and Magnitude Based Pruning for Sparse Deep Neural Networks
    Belay, Kaleab
    [J]. THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 13126 - 13127
  • [7] Magnitude and Uncertainty Pruning Criterion for Neural Networks
    Ko, Vinnie
    Oehmcke, Stefan
    Gieseke, Fabian
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2019, : 2317 - 2326
  • [8] Optimizing Neural Network Efficiency with Hybrid Magnitude-Based and Node Pruning for Energy-efficient Computing in IoT
    Uddin, Mohammad Helal
    Baidya, Sabur
    [J]. PROCEEDINGS 8TH ACM/IEEE CONFERENCE ON INTERNET OF THINGS DESIGN AND IMPLEMENTATION, IOTDI 2023, 2023, : 474 - 476
  • [9] Spectral Pruning for Recurrent Neural Networks
    Furuya, Takashi
    Suetake, Kazuma
    Taniguchi, Koichi
    Kusumoto, Hiroyuki
    Saiin, Ryuji
    Daimon, Tomohiro
    [J]. INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 151, 2022, 151
  • [10] Global Optimization of the Stage-wise Superstructure Model for Heat Exchanger Networks
    Faria, Debora C.
    Kim, Sung Young
    Bagajewicz, Miguel J.
    [J]. INDUSTRIAL & ENGINEERING CHEMISTRY RESEARCH, 2015, 54 (05) : 1595 - 1604