Model pruning based on filter similarity for edge device deployment

被引:0
|
作者
Wu, Tingting [1 ,2 ,3 ,4 ]
Song, Chunhe [1 ,2 ,3 ]
Zeng, Peng [1 ,2 ,3 ]
机构
[1] Chinese Acad Sci, Shenyang Inst Automat, State Key Lab Robot, Shenyang, Peoples R China
[2] Chinese Acad Sci, Key Lab Networked Control Syst, Shenyang, Peoples R China
[3] Chinese Acad Sci, Inst Robot & Intelligent Mfg, Shenyang, Peoples R China
[4] Univ Chinese Acad Sci, Beijing, Peoples R China
基金
国家重点研发计划;
关键词
network acceleration; filter pruning; edge intelligence; network compression; convolutional neural networks;
D O I
10.3389/fnbot.2023.1132679
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Filter pruning is widely used for inference acceleration and compatibility with off-the-shelf hardware devices. Some filter pruning methods have proposed various criteria to approximate the importance of filters, and then sort the filters globally or locally to prune the redundant parameters. However, the current criterion-based methods have problems: (1) parameters with smaller criterion values for extracting edge features are easily ignored, and (2) there is a strong correlation between different criteria, resulting in similar pruning structures. In this article, we propose a novel simple but effective pruning method based on filter similarity, which is used to evaluate the similarity between filters instead of the importance of a single filter. The proposed method first calculates the similarity of the filters pairwise in one convolutional layer and then obtains the similarity distribution. Finally, the filters with high similarity to others are deleted from the distribution or set to zero. In addition, the proposed algorithm does not need to specify the pruning rate for each layer, and only needs to set the desired FLOPs or parameter reduction to obtain the final compression model. We also provide iterative pruning strategies for hard pruning and soft pruning to satisfy the tradeoff requirements of accuracy and memory in different scenarios. Extensive experiments on various representative benchmark datasets across different network architectures demonstrate the effectiveness of our proposed method. For example, on CIFAR10, the proposed algorithm achieves 61.1% FLOPs reduction by removing 58.3% of the parameters, with no loss in Top-1 accuracy on ResNet-56; and reduces 53.05% FLOPs on ResNet-50 with only 0.29% Top-1 accuracy degradation on ILSVRC-2012.
引用
收藏
页数:16
相关论文
共 50 条
  • [1] Filter Pruning Method for Inference Time Acceleration Based on YOLOX in Edge Device
    Jeon, Jihun
    Kang, Jin-Ku
    Kim, Yongwoo
    [J]. 2022 19TH INTERNATIONAL SOC DESIGN CONFERENCE (ISOCC), 2022, : 354 - 355
  • [2] QSFM: Model Pruning Based on Quantified Similarity Between Feature Maps for AI on Edge
    Wang, Zidu
    Liu, Xuexin
    Huang, Long
    Chen, Yunqing
    Zhang, Yufei
    Lin, Zhikang
    Wang, Rui
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (23) : 24506 - 24515
  • [3] Similarity Based Filter Pruning for Efficient Super-Resolution Models
    Chu, Chu
    Chen, Li
    Gao, Zhiyong
    [J]. 2020 IEEE INTERNATIONAL SYMPOSIUM ON BROADBAND MULTIMEDIA SYSTEMS AND BROADCASTING (BMSB), 2020,
  • [4] SFP: Similarity-based filter pruning for deep neural networks
    Li, Guoqing
    Li, Rengang
    Li, Tuo
    Shen, Chaoyao
    Zou, Xiaofeng
    Wang, Jiuyang
    Wang, Changhong
    Li, Nanjun
    [J]. INFORMATION SCIENCES, 2025, 689
  • [5] A Passive Similarity based CNN Filter Pruning for Efficient Acoustic Scene Classification
    Singh, Arshdeep
    Plumbley, Mark D.
    [J]. INTERSPEECH 2022, 2022, : 2433 - 2437
  • [6] Edge devices object detection by filter pruning
    Crescitelli, Viviana
    Miura, Seiji
    Ono, Goichi
    Kohmu, Naohiro
    [J]. 2021 26TH IEEE INTERNATIONAL CONFERENCE ON EMERGING TECHNOLOGIES AND FACTORY AUTOMATION (ETFA), 2021,
  • [7] Cluster Pruning: An Efficient Filter Pruning Method for Edge AI Vision Applications
    Gamanayake, Chinthaka
    Jayasinghe, Lahiru
    Ng, Benny Kai Kiat
    Yuen, Chau
    [J]. IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2020, 14 (04) : 802 - 816
  • [8] Magnitude and Similarity Based Variable Rate Filter Pruning for Efficient Convolution Neural Networks
    Ghimire, Deepak
    Kim, Seong-Heum
    [J]. APPLIED SCIENCES-BASEL, 2023, 13 (01):
  • [9] A SIMPLE HYBRID FILTER PRUNING FOR EFFICIENT EDGE INFERENCE
    Basha, S. H. Shabbeer
    Gowda, Sheethal N.
    Dakala, Jayachandra
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 3398 - 3402
  • [10] Joint Model Pruning and Device Selection for Communication-Efficient Federated Edge Learning
    Liu, Shengli
    Yu, Guanding
    Yin, Rui
    Yuan, Jiantao
    Shen, Lei
    Liu, Chonghe
    [J]. IEEE TRANSACTIONS ON COMMUNICATIONS, 2022, 70 (01) : 231 - 244