ON THE ROLE OF STRUCTURED PRUNING FOR NEURAL NETWORK COMPRESSION

被引:5
|
作者
Bragagnolo, Andrea [1 ,2 ]
Tartaglione, Enzo [1 ]
Fiandrotti, Attilio [1 ]
Grangetto, Marco [1 ]
机构
[1] Univ Turin, Comp Sci Dept, I-10149 Turin, Italy
[2] Synesthesia Srl, Turin, TO, Italy
关键词
Pruning; Deep learning; Compression; MPEG-7;
D O I
10.1109/ICIP42928.2021.9506708
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This works explores the benefits of structured parameter pruning in the framework of the MPEG standardization efforts for neural network compression. First less relevant parameters are pruned from the network, then remaining parameters are quantized and finally quantized parameters are entropy coded. We consider an unstructured pruning strategy that maximizes the number of pruned parameters at the price of randomly sparse tensors and a structured strategy that prunes fewer parameters yet yields regularly sparse tensors. We show that structured pruning enables better end-to-end compression despite lower pruning ratio because it boosts the efficiency of the arithmetic coder. As a bonus, once decompressed, the network memory footprint is lower as well as its inference time.
引用
收藏
页码:3527 / 3531
页数:5
相关论文
共 50 条
  • [1] Adversarial Structured Neural Network Pruning
    Cai, Xingyu
    Yi, Jinfeng
    Zhang, Fan
    Rajasekaran, Sanguthevar
    [J]. PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT (CIKM '19), 2019, : 2433 - 2436
  • [2] Dirichlet Pruning for Neural Network Compression
    Adamczewski, Kamil
    Park, Mijung
    [J]. 24TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS (AISTATS), 2021, 130
  • [3] ScoringNet: A Neural Network Based Pruning Criteria for Structured Pruning
    Wang S.
    Zhang Z.
    [J]. Scientific Programming, 2023, 2023
  • [4] Automated Pruning for Deep Neural Network Compression
    Manessi, Franco
    Rozza, Alessandro
    Bianco, Simone
    Napoletano, Paolo
    Schettini, Raimondo
    [J]. 2018 24TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2018, : 657 - 664
  • [5] Quantisation and Pruning for Neural Network Compression and Regularisation
    Paupamah, Kimessha
    James, Steven
    Klein, Richard
    [J]. 2020 INTERNATIONAL SAUPEC/ROBMECH/PRASA CONFERENCE, 2020, : 295 - 300
  • [6] Neural Network Compression and Acceleration by Federated Pruning
    Pei, Songwen
    Wu, Yusheng
    Qiu, Meikang
    [J]. ALGORITHMS AND ARCHITECTURES FOR PARALLEL PROCESSING, ICA3PP 2020, PT II, 2020, 12453 : 173 - 183
  • [7] Structured pruning via feature channels similarity and mutual learning for convolutional neural network compression
    Wei Yang
    Yancai Xiao
    [J]. Applied Intelligence, 2022, 52 : 14560 - 14570
  • [8] Structured pruning via feature channels similarity and mutual learning for convolutional neural network compression
    Yang, Wei
    Xiao, Yancai
    [J]. APPLIED INTELLIGENCE, 2022, 52 (12) : 14560 - 14570
  • [9] PRUNE YOUR NEURONS BLINDLY: NEURAL NETWORK COMPRESSION THROUGH STRUCTURED CLASS-BLIND PRUNING
    Salama, Abdullah
    Ostapenko, Oleksiy
    Klein, Tassilo
    Nabi, Moin
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 2802 - 2806
  • [10] Revisiting Random Channel Pruning for Neural Network Compression
    Li, Yawei
    Adamczewski, Kamil
    Li, Wen
    Gu, Shuhang
    Timofte, Radu
    Van Gool, Luc
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 191 - 201