Exploiting Variable Precision Computation Array for Scalable Neural Network Accelerators

被引:0
|
作者
Yang, Shaofei [1 ]
Liu, Longjun [1 ]
Li, Baoting [1 ]
Sun, Hongbin [1 ]
Zheng, Nanning [1 ]
机构
[1] Xi An Jiao Tong Univ, Inst Artificial Intelligence & Robot, Xian 710049, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep Neural Networks; Accelerator; Energy Efficiency Computing Array; Dynamic Quantization; Resiliency;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we present a flexible Variable Precision Computation Array (VPCA) component for different accelerators, which leverages a sparsification scheme for activations and a low bits serial-parallel combination computation unit for improving the efficiency and resiliency of accelerators. The VPCA can dynamically decompose the width of activation/weights (from 32bit to 3bit in different accelerators) into 2-bits serial computation units while the 2bits computing units can be combined in parallel computing for high throughput. We propose an on-the-fly compressing and calculating strategy SLE-CLC (single lane encoding, cross lane calculation), which could further improve performance of 2-bit parallel computing. The experiments results on image classification datasets show VPCA can outperforms DaDianNao, Stripes, Loom-2bit by 4.67x, 2.42x, 1.52x without other overhead on convolution layers.
引用
收藏
页码:315 / 319
页数:5
相关论文
共 50 条
  • [41] Survey of Precision-Scalable Multiply-Accumulate Units for Neural-Network Processing
    Camus, Vincent
    Enz, Christian
    Verhelst, Marian
    2019 IEEE INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE CIRCUITS AND SYSTEMS (AICAS 2019), 2019, : 57 - 61
  • [42] Review of ASIC accelerators for deep neural network
    Machupalli, Raju
    Hossain, Masum
    Mandal, Mrinal
    MICROPROCESSORS AND MICROSYSTEMS, 2022, 89
  • [43] Initial study of reconfigurable neural network accelerators
    1600, Institute of Electrical and Electronics Engineers Inc., United States
  • [44] Memory Trojan Attack on Neural Network Accelerators
    Zhao, Yang
    Hu, Xing
    Li, Shuangchen
    Ye, Jing
    Deng, Lei
    Ji, Yu
    Xu, Jianyu
    Wu, Dong
    Xie, Yuan
    2019 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE), 2019, : 1415 - 1420
  • [45] Initial Study of Reconfigurable Neural Network Accelerators
    Ohba, Momoka
    Shindo, Satoshi
    Miwa, Shinobu
    Tsumura, Tomoaki
    Yamaki, Hayato
    Honda, Hiroki
    2016 FOURTH INTERNATIONAL SYMPOSIUM ON COMPUTING AND NETWORKING (CANDAR), 2016, : 707 - 709
  • [46] Partition and Scheduling Algorithms for Neural Network Accelerators
    Chen, Xiaobing
    Peng, Shaohui
    Jin, Luyang
    Zhuang, Yimin
    Song, Jin
    Du, Weijian
    Liu, Shaoli
    Zhi, Tian
    ADVANCED PARALLEL PROCESSING TECHNOLOGIES (APPT 2019), 2019, 11719 : 55 - 67
  • [47] Approximate Adders for Deep Neural Network Accelerators
    Raghuram, S.
    Shashank, N.
    2022 35TH INTERNATIONAL CONFERENCE ON VLSI DESIGN (VLSID 2022) HELD CONCURRENTLY WITH 2022 21ST INTERNATIONAL CONFERENCE ON EMBEDDED SYSTEMS (ES 2022), 2022, : 210 - 215
  • [48] Photonic Computing and Communication for Neural Network Accelerators
    Xia, Chengpeng
    Chen, Yawen
    Zhang, Haibo
    Zhang, Hao
    Wu, Jigang
    PARALLEL AND DISTRIBUTED COMPUTING, APPLICATIONS AND TECHNOLOGIES, PDCAT 2021, 2022, 13148 : 121 - 128
  • [49] Making Memristive Neural Network Accelerators Reliable
    Feinberg, Ben
    Wang, Shibo
    Ipek, Engin
    2018 24TH IEEE INTERNATIONAL SYMPOSIUM ON HIGH PERFORMANCE COMPUTER ARCHITECTURE (HPCA), 2018, : 52 - 65
  • [50] A comprehensive survey on graph neural network accelerators
    Liu, Jingyu
    Chen, Shi
    Shen, Li
    FRONTIERS OF COMPUTER SCIENCE, 2025, 19 (02)