Structured pruning of recurrent neural networks through neuron selection

被引:26
|
作者
Wen, Liangjian [1 ]
Zhang, Xuanyang [1 ]
Bai, Haoli [2 ]
Xu, Zenglin [1 ,3 ]
机构
[1] Univ Elect Sci & Technol China, Sch Comp Sci & Engn, SMILE Lab, Chengdu 610031, Peoples R China
[2] Chinese Univ Hong Kong, Dept Comp Sci & Engn, Shatin, Hong Kong 999077, Peoples R China
[3] Ctr Artificial Intelligence, Peng Cheng Lab, Shenzhen, Guangdong, Peoples R China
关键词
Feature selection; Recurrent neural networks; Learning sparse models; Model compression;
D O I
10.1016/j.neunet.2019.11.018
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recurrent neural networks (RNNs) have recently achieved remarkable successes in a number of applications. However, the huge sizes and computational burden of these models make it difficult for their deployment on edge devices. A practically effective approach is to reduce the overall storage and computation costs of RNNs by network pruning techniques. Despite their successful applications, those pruning methods based on Lasso either produce irregular sparse patterns in weight matrices, which is not helpful in practical speedup. To address these issues, we propose a structured pruning method through neuron selection which can remove the independent neuron of RNNs. More specifically, we introduce two sets of binary random variables, which can be interpreted as gates or switches to the input neurons and the hidden neurons, respectively. We demonstrate that the corresponding optimization problem can be addressed by minimizing the L-0 norm of the weight matrix. Finally, experimental results on language modeling and machine reading comprehension tasks have indicated the advantages of the proposed method in comparison with state-of-the-art pruning competitors. In particular, nearly 20x practical speedup during inference was achieved without losing performance for the language model on the Penn TreeBank dataset, indicating the promising performance of the proposed method. (C) 2019 Elsevier Ltd. All rights reserved.
引用
收藏
页码:134 / 141
页数:8
相关论文
共 50 条
  • [1] Spectral Pruning for Recurrent Neural Networks
    Furuya, Takashi
    Suetake, Kazuma
    Taniguchi, Koichi
    Kusumoto, Hiroyuki
    Saiin, Ryuji
    Daimon, Tomohiro
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 151, 2022, 151
  • [2] Structured Pruning of Deep Convolutional Neural Networks
    Anwar, Sajid
    Hwang, Kyuyeon
    Sung, Wonyong
    ACM JOURNAL ON EMERGING TECHNOLOGIES IN COMPUTING SYSTEMS, 2017, 13 (03)
  • [3] Structured pruning of neural networks for constraints learning
    Cacciola, Matteo
    Frangioni, Antonio
    Lodi, Andrea
    Operations Research Letters, 2024, 57
  • [4] Leveraging Structured Pruning of Convolutional Neural Networks
    Tessier, Hugo
    Gripon, Vincent
    Leonardon, Mathieu
    Arzel, Matthieu
    Bertrand, David
    Hannagan, Thomas
    2022 IEEE WORKSHOP ON SIGNAL PROCESSING SYSTEMS (SIPS), 2022, : 174 - 179
  • [5] Automatic Pruning Rate Derivation for Structured Pruning of Deep Neural Networks
    Sakai, Yasufumi
    Iwakawa, Akinori
    Tabaru, Tsuguchika
    Inoue, Atsuki
    Kawaguchi, Hiroshi
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 2561 - 2567
  • [6] Structured Pruning for Deep Convolutional Neural Networks: A Survey
    He, Yang
    Xiao, Lingao
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (05) : 2900 - 2919
  • [7] Pruning Approaches for Selection of Neural Networks Structure
    Abid, Slim
    Chtourou, Mohamed
    Djemel, Mohamed
    2013 10TH INTERNATIONAL MULTI-CONFERENCE ON SYSTEMS, SIGNALS & DEVICES (SSD), 2013,
  • [8] PRUNING RECURRENT NEURAL NETWORKS FOR IMPROVED GENERALIZATION PERFORMANCE
    GILES, CL
    OMLIN, CW
    IEEE TRANSACTIONS ON NEURAL NETWORKS, 1994, 5 (05): : 848 - 851
  • [9] PeakRNN and StatsRNN: Dynamic Pruning in Recurrent Neural Networks
    Jelcicova, Zuzana
    Jones, Rasmus
    Blix, David Thorn
    Verhelst, Marian
    Sparso, Jens
    29TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO 2021), 2021, : 416 - 420
  • [10] Structured Term Pruning for Computational Efficient Neural Networks Inference
    Huang, Kai
    Li, Bowen
    Chen, Siang
    Claesen, Luc
    Xi, Wei
    Chen, Junjian
    Jiang, Xiaowen
    Liu, Zhili
    Xiong, Dongliang
    Yan, Xiaolang
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2023, 42 (01) : 190 - 203