Tetris: Re-architecting Convolutional Neural Network Computation for Machine Learning Accelerators

被引:29
|
作者
Lu, Hang [1 ,2 ]
Wei, Xin [2 ]
Lin, Ning [2 ]
Yan, Guihai [1 ,2 ]
Li, Xiao-Wei [1 ,2 ]
机构
[1] Chinese Acad Sci, Inst Comp Technol, State Key Lab Comp Architecture, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1145/3240765.3240855
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Inference efficiency is the predominant consideration in designing deep learning accelerators. Previous work mainly focuses on skipping zero values to deal with remarkable ineffectual computation, while zero bits in non-zero values, as another major source of ineffectual computation, is often ignored. The reason lies on the difficulty of extracting essential bits during operating multiply-and-accumulate (MAC) in the processing element. Based on the fact that zero bits occupy as high as 68.9% fraction in the overall weights of modern deep convolutional neural network models, this paper firstly proposes a weight kneading technique that could eliminate ineffectual computation caused by either zero value weights or zero bits in non-zero weights, simultaneously. Besides, a split-and-accumulate (SAC) computing pattern in replacement of conventional MAC, as well as the corresponding hardware accelerator design called Tetris are proposed to support weight kneading at the hardware level. Experimental results prove that Tetris could speed up inference up to 1.50x, and improve power efficiency up to 5.33x compared with the state-of-the-art baselines.
引用
收藏
页数:8
相关论文
共 50 条
  • [21] Latency-Insensitive Controller for Convolutional Neural Network Accelerators
    Seo, Youngho
    Lee, Sanghun
    Kim, Sunwoo
    Wang, Jooho
    Park, Sungkyung
    Park, Chester Sungchung
    2019 INTERNATIONAL SOC DESIGN CONFERENCE (ISOCC), 2019, : 249 - 250
  • [22] Optimizing Memory Efficiency for Deep Convolutional Neural Network Accelerators
    Li, Xiaowei
    Li, Jiajun
    Yan, Guihai
    JOURNAL OF LOW POWER ELECTRONICS, 2018, 14 (04) : 496 - 507
  • [23] Robust visual tracking based on convolutional neural network with extreme learning machine
    Sun, Rui
    Wang, Xu
    Yan, Xiaoxing
    MULTIMEDIA TOOLS AND APPLICATIONS, 2019, 78 (06) : 7543 - 7562
  • [24] Robust visual tracking based on convolutional neural network with extreme learning machine
    Rui Sun
    Xu Wang
    Xiaoxing Yan
    Multimedia Tools and Applications, 2019, 78 : 7543 - 7562
  • [25] Efficiency of corporate debt financing based on machine learning and convolutional neural network
    Zhao, Jing
    MICROPROCESSORS AND MICROSYSTEMS, 2021, 83
  • [26] An effective classifier based on convolutional neural network and regularized extreme learning machine
    He, Chunmei
    Kang, Hongyu
    Yao, Tong
    Li, Xiaorui
    MATHEMATICAL BIOSCIENCES AND ENGINEERING, 2019, 16 (06) : 8309 - 8321
  • [27] Convolutional neural network extreme learning machine for effective classification of hyperspectral images
    Cao, Faxian
    Yang, Zhijing
    Ren, Jinchang
    Ling, Bingo Wing-Kuen
    JOURNAL OF APPLIED REMOTE SENSING, 2018, 12 (03):
  • [28] ConvELM: Exploiting Extreme Learning Machine on Convolutional Neural Network for Age Estimation
    Apuandi, Ismar
    Rachmawati, Ema
    Kosala, Gamma
    2023 INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE IN INFORMATION AND COMMUNICATION, ICAIIC, 2023, : 407 - 412
  • [29] Facial Expressions Recognition through Convolutional Neural Network and Extreme Learning Machine
    Jammoussi, Imen
    Ben Nasr, Mounir
    Chtourou, Mohamed
    PROCEEDINGS OF THE 2020 17TH INTERNATIONAL MULTI-CONFERENCE ON SYSTEMS, SIGNALS & DEVICES (SSD 2020), 2020, : 162 - 166
  • [30] Recognition of industrial machine parts based on transfer learning with convolutional neural network
    Li, Qiaoyang
    Chen, Guiming
    PLOS ONE, 2021, 16 (01):