Structured pruning of recurrent neural networks through neuron selection

被引:26
|
作者
Wen, Liangjian [1 ]
Zhang, Xuanyang [1 ]
Bai, Haoli [2 ]
Xu, Zenglin [1 ,3 ]
机构
[1] Univ Elect Sci & Technol China, Sch Comp Sci & Engn, SMILE Lab, Chengdu 610031, Peoples R China
[2] Chinese Univ Hong Kong, Dept Comp Sci & Engn, Shatin, Hong Kong 999077, Peoples R China
[3] Ctr Artificial Intelligence, Peng Cheng Lab, Shenzhen, Guangdong, Peoples R China
关键词
Feature selection; Recurrent neural networks; Learning sparse models; Model compression;
D O I
10.1016/j.neunet.2019.11.018
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recurrent neural networks (RNNs) have recently achieved remarkable successes in a number of applications. However, the huge sizes and computational burden of these models make it difficult for their deployment on edge devices. A practically effective approach is to reduce the overall storage and computation costs of RNNs by network pruning techniques. Despite their successful applications, those pruning methods based on Lasso either produce irregular sparse patterns in weight matrices, which is not helpful in practical speedup. To address these issues, we propose a structured pruning method through neuron selection which can remove the independent neuron of RNNs. More specifically, we introduce two sets of binary random variables, which can be interpreted as gates or switches to the input neurons and the hidden neurons, respectively. We demonstrate that the corresponding optimization problem can be addressed by minimizing the L-0 norm of the weight matrix. Finally, experimental results on language modeling and machine reading comprehension tasks have indicated the advantages of the proposed method in comparison with state-of-the-art pruning competitors. In particular, nearly 20x practical speedup during inference was achieved without losing performance for the language model on the Penn TreeBank dataset, indicating the promising performance of the proposed method. (C) 2019 Elsevier Ltd. All rights reserved.
引用
收藏
页码:134 / 141
页数:8
相关论文
共 50 条
  • [31] Structured Pruning for Deep Neural Networks with Adaptive Pruning Rate Derivation Based on Connection Sensitivity and Loss Function
    Sakai, Yasufumi
    Eto, Yu
    Teranishi, Yuta
    JOURNAL OF ADVANCES IN INFORMATION TECHNOLOGY, 2022, 13 (03) : 295 - 300
  • [32] Implicit neural representation steganography by neuron pruning
    Dong, Weina
    Liu, Jia
    Chen, Lifeng
    Sun, Wenquan
    Pan, Xiaozhong
    Ke, Yan
    MULTIMEDIA SYSTEMS, 2024, 30 (05)
  • [33] Safe Crossover of Neural Networks Through Neuron Alignment
    Uriot, Thomas
    Izzo, Dario
    GECCO'20: PROCEEDINGS OF THE 2020 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE, 2020, : 435 - 443
  • [34] Optimal Neuron Selection and Generalization: NK Ensemble Neural Networks
    Whitley, Darrell
    Tinos, Renato
    Chicano, Francisco
    PARALLEL PROBLEM SOLVING FROM NATURE - PPSN XV, PT II, 2018, 11102 : 449 - 460
  • [35] Sequence Labelling in Structured Domains with Hierarchical Recurrent Neural Networks
    Fernandez, Santiago
    Graves, Alex
    Schmidhuber, Juergen
    20TH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2007, : 774 - 779
  • [36] RECURRENT CONVOLUTIONAL NEURAL NETWORKS FOR STRUCTURED SPEECH ACT TAGGING
    Ushio, Takashi
    Shi, Hongjie
    Endo, Mitsuru
    Yamagami, Katsuyoshi
    Horii, Noriaki
    2016 IEEE WORKSHOP ON SPOKEN LANGUAGE TECHNOLOGY (SLT 2016), 2016, : 518 - 524
  • [37] Unsupervised Neuron Selection for Mitigating Catastrophic Forgetting in Neural Networks
    Goodrich, Ben
    Arel, Itamar
    2014 IEEE 57TH INTERNATIONAL MIDWEST SYMPOSIUM ON CIRCUITS AND SYSTEMS (MWSCAS), 2014, : 997 - 1000
  • [38] Online identification of a mechatronic system with structured recurrent neural networks
    Hintz, C
    Angerer, B
    Schröder, D
    ISIE 2002: PROCEEDINGS OF THE 2002 IEEE INTERNATIONAL SYMPOSIUM ON INDUSTRIAL ELECTRONICS, VOLS 1-4, 2002, : 288 - 293
  • [39] Extended Kalman filter-based pruning method for recurrent neural networks
    Sum, J
    Chan, LW
    Leung, CS
    Young, GH
    NEURAL COMPUTATION, 1998, 10 (06) : 1481 - 1505
  • [40] Stage-Wise Magnitude-Based Pruning for Recurrent Neural Networks
    Li, Guiying
    Yang, Peng
    Qian, Chao
    Hong, Richang
    Tang, Ke
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (02) : 1666 - 1680