Structured pruning of recurrent neural networks through neuron selection

被引:26
|
作者
Wen, Liangjian [1 ]
Zhang, Xuanyang [1 ]
Bai, Haoli [2 ]
Xu, Zenglin [1 ,3 ]
机构
[1] Univ Elect Sci & Technol China, Sch Comp Sci & Engn, SMILE Lab, Chengdu 610031, Peoples R China
[2] Chinese Univ Hong Kong, Dept Comp Sci & Engn, Shatin, Hong Kong 999077, Peoples R China
[3] Ctr Artificial Intelligence, Peng Cheng Lab, Shenzhen, Guangdong, Peoples R China
关键词
Feature selection; Recurrent neural networks; Learning sparse models; Model compression;
D O I
10.1016/j.neunet.2019.11.018
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recurrent neural networks (RNNs) have recently achieved remarkable successes in a number of applications. However, the huge sizes and computational burden of these models make it difficult for their deployment on edge devices. A practically effective approach is to reduce the overall storage and computation costs of RNNs by network pruning techniques. Despite their successful applications, those pruning methods based on Lasso either produce irregular sparse patterns in weight matrices, which is not helpful in practical speedup. To address these issues, we propose a structured pruning method through neuron selection which can remove the independent neuron of RNNs. More specifically, we introduce two sets of binary random variables, which can be interpreted as gates or switches to the input neurons and the hidden neurons, respectively. We demonstrate that the corresponding optimization problem can be addressed by minimizing the L-0 norm of the weight matrix. Finally, experimental results on language modeling and machine reading comprehension tasks have indicated the advantages of the proposed method in comparison with state-of-the-art pruning competitors. In particular, nearly 20x practical speedup during inference was achieved without losing performance for the language model on the Penn TreeBank dataset, indicating the promising performance of the proposed method. (C) 2019 Elsevier Ltd. All rights reserved.
引用
收藏
页码:134 / 141
页数:8
相关论文
共 50 条
  • [21] Diverse Neuron Type Selection for Convolutional Neural Networks
    Zhu, Guibo
    Zhang, Zhaoxiang
    Zhang, Xu-Yao
    Liu, Cheng-Lin
    PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 3560 - 3566
  • [22] Sparseness Ratio Allocation and Neuron Re-pruning for Neural Networks Compression
    Guo, Li
    Zhou, Dajiang
    Zhou, Jinjia
    Kimura, Shinji
    2018 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2018,
  • [23] ITERATIVE PRUNING IN 2ND-ORDER RECURRENT NEURAL NETWORKS
    CASTELLANO, G
    FANELLI, AM
    PELILLO, M
    NEURAL PROCESSING LETTERS, 1995, 2 (06) : 5 - 8
  • [24] Exploring Compute-in-Memory Architecture Granularity for Structured Pruning of Neural Networks
    Meng, Fan-Hsuan
    Wang, Xinxin
    Wang, Ziyu
    Lee, Eric Yeu-Jer
    Lu, Wei D.
    IEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS, 2022, 12 (04) : 858 - 866
  • [25] Adversarial Structured Neural Network Pruning
    Cai, Xingyu
    Yi, Jinfeng
    Zhang, Fan
    Rajasekaran, Sanguthevar
    PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT (CIKM '19), 2019, : 2433 - 2436
  • [26] Structured Pruning for Deep Convolutional Neural Networks via Adaptive Sparsity Regularization
    Shao, Tuanjie
    Shin, Dongkun
    2022 IEEE 46TH ANNUAL COMPUTERS, SOFTWARE, AND APPLICATIONS CONFERENCE (COMPSAC 2022), 2022, : 982 - 987
  • [27] Structured Pruning of Convolutional Neural Networks via L1 Regularization
    Yang, Chen
    Yang, Zhenghong
    Khattak, Abdul Mateen
    Yang, Liu
    Zhang, Wenxin
    Gao, Wanlin
    Wang, Minjuan
    IEEE ACCESS, 2019, 7 : 106385 - 106394
  • [28] ScoringNet: A Neural Network Based Pruning Criteria for Structured Pruning
    Wang S.
    Zhang Z.
    Scientific Programming, 2023, 2023
  • [29] Pruning of Deep Spiking Neural Networks through Gradient Rewiring
    Chen, Yanqi
    Yu, Zhaofei
    Fang, Wei
    Huang, Tiejun
    Tian, Yonghong
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 1713 - 1721
  • [30] Efficient variable selection batch pruning algorithm for artificial neural networks
    Kovalishyn, Vasyl
    Poda, Gennady
    CHEMOMETRICS AND INTELLIGENT LABORATORY SYSTEMS, 2015, 149 : 10 - 16