Structured Sparse Ternary Weight Coding of Deep Neural Networks for Efficient Hardware Implementations

被引:0
|
作者
Boo, Yoonho [1 ]
Sung, Wonyong [1 ]
机构
[1] Seoul Natl Univ, Dept Elect Engn & Comp Sci, Seoul 151744, South Korea
基金
新加坡国家研究基金会;
关键词
Deep neural networks; weight storage compression; structured sparsity; fixed-point quantization; network pruning;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNs) usually demand a large amount of operations for real-time inference. Especially, fully-connected layers contain a large number of weights, thus they usually need many off-chip memory accesses for inference. We propose a weight compression method for deep neural networks, which allows values of + 1 or -1 only at predetermined positions of the weights so that decoding using a table can be conducted easily. For example, the structured sparse (8,2) coding allows at most two non-zero values among eight weights. This method not only enables multiplication-free DNN implementations but also compresses the weight storage by up to x32 compared to floating-point networks. Weight distribution normalization and gradual pruning techniques are applied to mitigate the performance degradation. The experiments are conducted with fully-connected deep neural networks and convolutional neural networks.
引用
收藏
页数:6
相关论文
共 50 条
  • [31] Spiking Neural Networks - Algorithms, Hardware Implementations and Applications
    Kulkarni, Shruti R.
    Babu, Anakha V.
    Rajendran, Bipin
    2017 IEEE 60TH INTERNATIONAL MIDWEST SYMPOSIUM ON CIRCUITS AND SYSTEMS (MWSCAS), 2017, : 426 - 431
  • [32] A Review of Algorithms and Hardware Implementations for Spiking Neural Networks
    Duy-Anh Nguyen
    Xuan-Tu Tran
    Iacopi, Francesca
    JOURNAL OF LOW POWER ELECTRONICS AND APPLICATIONS, 2021, 11 (02)
  • [33] Quantized rewiring: hardware-aware training of sparse deep neural networks
    Petschenig, Horst
    Legenstein, Robert
    NEUROMORPHIC COMPUTING AND ENGINEERING, 2023, 3 (02):
  • [34] Spiking Neural Networks Hardware Implementations and Challenges: A Survey
    Bouvier, Maxence
    Valentian, Alexandre
    Mesquida, Thomas
    Rummens, Francois
    Reyboz, Marina
    Vianello, Elisa
    Beigne, Edith
    ACM JOURNAL ON EMERGING TECHNOLOGIES IN COMPUTING SYSTEMS, 2019, 15 (02)
  • [35] SUPERVISED DEEP SPARSE CODING NETWORKS
    Sun, Xiaoxia
    Nasrabadi, Nasser M.
    Tran, Trac D.
    2018 25TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2018, : 346 - 350
  • [36] BAYESIAN NEURAL NETWORKS FOR SPARSE CODING
    Kuzin, Danil
    Isupova, Olga
    Mihaylova, Lyudmila
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 2992 - 2996
  • [37] Sparse coding for layered neural networks
    Katayama, K
    Sakata, Y
    Horiguchi, T
    PHYSICA A-STATISTICAL MECHANICS AND ITS APPLICATIONS, 2002, 310 (3-4) : 532 - 546
  • [38] Hardware implementations of neural networks and the Random Neural Network Chip (RNNC)
    Aybay, I
    Çerkez, C
    Halici, U
    Badaroglu, M
    ADVANCES IN COMPUTER AND INFORMATION SCIENCES '98, 1998, 53 : 157 - 161
  • [39] Harmonious Coexistence of Structured Weight Pruning and Ternarization for Deep Neural Networks
    Yang, Li
    He, Zhezhi
    Fan, Deliang
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 6623 - 6630
  • [40] Sparse coding of pathology slides compared to transfer learning with deep neural networks
    Will Fischer
    Sanketh S. Moudgalya
    Judith D. Cohn
    Nga T. T. Nguyen
    Garrett T. Kenyon
    BMC Bioinformatics, 19