Sparse Ternary Connect: Convolutional Neural Networks Using Ternarized Weights with Enhanced Sparsity

被引:0
|
作者
Jin, Canran [1 ]
Sun, Heming [1 ]
Kimura, Shinji [1 ]
机构
[1] Waseda Univ, Grad Sch Informat Prod & Syst, Wakamatsu Ku, 2-7 Hibikino, Kitakyushu, Fukuoka 8080135, Japan
关键词
D O I
暂无
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Convolutional Neural Networks (CNNs) are indispensable in a wide range of tasks to achieve state-of-the-art results. In this work, we exploit ternary weights in both inference and training of CNNs and further propose Sparse Ternary Connect (STC) where kernel weights in float value are converted to 1, -1 and 0 based on a new conversion rule with the controlled ratio of 0. STC can save hardware resource a lot with small degradation of precision. The experimental evaluation on 2 popular datasets (CIFAR-10 and SVHN) shows that the proposed method can reduce resource utilization (by 28.9% of LUT, 25.3% of FF, 97.5% of DSP and 88.7% of BRAM on Xilinx Kintex-7 FPGA) with less than 0.5% accuracy loss.
引用
收藏
页码:190 / 195
页数:6
相关论文
共 50 条
  • [21] Impact of random weights on nonlinear system identification using convolutional neural networks
    Yu, Wen
    Pacheco, Mario
    INFORMATION SCIENCES, 2019, 477 : 1 - 14
  • [22] Accelerating Convolutional Neural Networks by Exploiting the Sparsity of Output Activation
    Fan, Zhihua
    Li, Wenming
    Wang, Zhen
    Liu, Tianyu
    Wu, Haibin
    Liu, Yanhuan
    Wu, Meng
    Wu, Xinxin
    Ye, Xiaochun
    Fan, Dongrui
    Sun, Ninghui
    An, Xuejun
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2023, 34 (12) : 3253 - 3265
  • [23] Compressing Low Precision Deep Neural Networks Using Sparsity-Induced Regularization in Ternary Networks
    Faraone, Julian
    Fraser, Nicholas
    Gambardella, Giulio
    Blott, Michaela
    Leong, Philip H. W.
    NEURAL INFORMATION PROCESSING (ICONIP 2017), PT II, 2017, 10635 : 393 - 404
  • [24] A Convolutional Spiking Neural Network Accelerator with the Sparsity-aware Memory and Compressed Weights
    Liu, Hanqing
    Cui, Xiaole
    Zhang, Sunrui
    Yin, Mingqi
    Jiang, Yuanyuan
    Cui, Xiaoxin
    2024 IEEE 35TH INTERNATIONAL CONFERENCE ON APPLICATION-SPECIFIC SYSTEMS, ARCHITECTURES AND PROCESSORS, ASAP 2024, 2024, : 163 - 171
  • [25] Radar Object Detection on a Vector Processor Using Sparse Convolutional Neural Networks
    Koehler, Daniel
    Meinl, Frank
    Blume, Holger
    EMBEDDED COMPUTER SYSTEMS: ARCHITECTURES, MODELING, AND SIMULATION, SAMOS 2024, PT I, 2025, 15226 : 138 - 154
  • [26] AGNet: Attention Guided Sparse Depth Completion Using Convolutional Neural Networks
    Liang, Xiaolong
    Jung, Cheolkon
    IEEE ACCESS, 2022, 10 : 10514 - 10522
  • [27] Accelerating the Inference Phase in Ternary Convolutional Neural Networks using Configurable Processors
    Chidambaram, Sivakumar
    Riviello, Alexandre
    Langlois, J. M. Pierre
    David, Jean-Pierre
    2018 CONFERENCE ON DESIGN AND ARCHITECTURES FOR SIGNAL AND IMAGE PROCESSING (DASIP), 2018, : 94 - 99
  • [28] Classification of Informative Frames in Colonoscopy Videos Using Convolutional Neural Networks with Binarized Weights
    Akbari, Mojtaba
    Mohrekesh, Majid
    Rafiei, Shima
    Soroushmehr, S. M. Reza
    Karimi, Nader
    Samavi, Shadrokh
    Najarian, Kayvan
    2018 40TH ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY (EMBC), 2018, : 65 - 68
  • [29] Dynamic block sparse reparameterization of convolutional neural networks
    Vooturi, Dharma Teja
    Varma, Girish
    Kothapalli, Kishore
    Proceedings - 2019 International Conference on Computer Vision Workshop, ICCVW 2019, 2019, : 3046 - 3053
  • [30] SparTen: A Sparse Tensor Accelerator for Convolutional Neural Networks
    Gondimalla, Ashish
    Chesnut, Noah
    Thottethodi, Mithuna
    Vijaykumar, T. N.
    MICRO'52: THE 52ND ANNUAL IEEE/ACM INTERNATIONAL SYMPOSIUM ON MICROARCHITECTURE, 2019, : 151 - 165