Exploring the Granularity of Sparsity in Convolutional Neural Networks

被引:312
|
作者
Mao, Huizi [1 ]
Han, Song [1 ]
Pool, Jeff [2 ]
Li, Wenshuo [3 ]
Liu, Xingyu [1 ]
Wang, Yu [3 ]
Dally, William J. [1 ,2 ]
机构
[1] Stanford Univ, Stanford, CA 94305 USA
[2] NVIDIA, Santa Clara, CA USA
[3] Tsinghua Univ, Beijing, Peoples R China
关键词
D O I
10.1109/CVPRW.2017.241
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Sparsity helps reducing the computation complexity of DNNs by skipping the multiplication with zeros. The granularity of sparsity affects the efficiency of hardware architecture and the prediction accuracy. In this paper we quantitatively measure the accuracy-sparsity relationship with different granularity. Coarse-grained sparsity brings more regular sparsity pattern, making it easier for hardware acceleration, and our experimental results show that coarse-grained sparsity have very small impact on the sparsity ratio given no loss of accuracy. Moreover, due to the index saving effect, coarse-grained sparsity is able to obtain similar or even better compression rates than fine-grained sparsity at the same accuracy threshold. Our analysis, which is based on the framework of a recent sparse convolutional neural network (SCNN) accelerator, further demonstrates that it saves 30% - 35% of memory references compared with fine-grained sparsity.
引用
收藏
页码:1927 / 1934
页数:8
相关论文
共 50 条
  • [1] Exploring Fine-Grained Sparsity in Convolutional Neural Networks for Efficient Inference
    Wang, Longguang
    Guo, Yulan
    Dong, Xiaoyu
    Wang, Yingqian
    Ying, Xinyi
    Lin, Zaiping
    An, Wei
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (04) : 4474 - 4493
  • [2] On Implicit Filter Level Sparsity in Convolutional Neural Networks
    Mehta, Dushyant
    Kim, Kwang In
    Theobalt, Christian
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 520 - 528
  • [3] Accelerating Convolutional Neural Networks by Exploiting the Sparsity of Output Activation
    Fan, Zhihua
    Li, Wenming
    Wang, Zhen
    Liu, Tianyu
    Wu, Haibin
    Liu, Yanhuan
    Wu, Meng
    Wu, Xinxin
    Ye, Xiaochun
    Fan, Dongrui
    Sun, Ninghui
    An, Xuejun
    [J]. IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2023, 34 (12) : 3253 - 3265
  • [4] SparseTrain: Exploiting Dataflow Sparsity for Efficient Convolutional Neural Networks Training
    Dai, Pengcheng
    Yang, Jianlei
    Ye, Xucheng
    Cheng, Xingzhou
    Luo, Junyu
    Song, Linghao
    Chen, Yiran
    Zhao, Weisheng
    [J]. PROCEEDINGS OF THE 2020 57TH ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2020,
  • [5] Sparsity Enables Data and Energy Efficient Spiking Convolutional Neural Networks
    Bhatt, Varun
    Ganguly, Udayan
    [J]. ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2018, PT I, 2018, 11139 : 263 - 272
  • [6] Emergence of Shape Bias in Convolutional Neural Networks through Activation Sparsity
    Li, Tianqin
    Wen, Ziqi
    Li, Yangfan
    Lee, Tai Sing
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [7] Exploring Hidden Dimensions in Parallelizing Convolutional Neural Networks
    Jia, Zhihao
    Lin, Sina
    Qi, Charles R.
    Aiken, Alex
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [8] Exploring complex pattern formation with convolutional neural networks
    Scholz, Christian
    Scholz, Sandy
    [J]. AMERICAN JOURNAL OF PHYSICS, 2022, 90 (02) : 141 - 151
  • [9] FEATURE SPARSITY IN CONVOLUTIONAL NEURAL NETWORKS FOR SCENE CLASSIFICATION OF REMOTE SENSING IMAGE
    Huang, Wei
    Wang, Qi
    Li, Xuelong
    [J]. 2019 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM (IGARSS 2019), 2019, : 3017 - 3020
  • [10] Latent HyperNet: Exploring the Layers of Convolutional Neural Networks
    Jordao, Artur
    Kloss, Ricardo
    Schwartz, William Robson
    [J]. 2018 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2018,