Dynamic Precision Multiplier For Deep Neural Network Accelerators

被引:1
|
作者
Ding, Chen [1 ]
Yuxiang, Huan [1 ]
Zheng, Lirong [1 ]
Zou, Zhuo [1 ]
机构
[1] Fudan Univ, Sch Informat Sci & Technol, Shanghai, Peoples R China
关键词
dynamic precision multiplier; Booth algorithm; mixed partial product selection structure;
D O I
10.1109/SOCC49529.2020.9524752
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The application of dynamic precision multipliers in the deep neural network accelerators can greatly improve system's data processing capacity under same memory bandwidth limitation. This paper presents a Dynamic Precision Multiplier (DPM) for deep learning accelerators to adapt to light-weight deep learning models with varied precision. The proposed DPM adopts Booth algorithm and Wallace Adder Tree to support parallel computation of signed/unsigned one 16-bit, two 8-bit or four 4-bit at run time. The DPM is further optimized with simplified partial product selection logic and mixed partial product selection structure techniques, reducing power cost for energy-efficient edge computing. The DPM is evaluated in both FPGA and ASIC flow, and the results show that 4-bit mode consumes the least energy among the three modes at 1.34pJ/word. It also saves nearly 22.38% and 232.17% of the power consumption under 16-bit and 8-bit mode respectively when comparing with previous similar designs.
引用
收藏
页码:180 / 184
页数:5
相关论文
共 50 条
  • [41] Exploration and Generation of Efficient FPGA-based Deep Neural Network Accelerators
    Ali, Nermine
    Philippe, Jean-Marc
    Tain, Benoit
    Coussy, Philippe
    2021 IEEE WORKSHOP ON SIGNAL PROCESSING SYSTEMS (SIPS 2021), 2021, : 123 - 128
  • [42] High-level Modeling of Manufacturing Faults in Deep Neural Network Accelerators
    Kundu, Shamik
    Soyyigit, Ahmet
    Hoque, Khaza Anuarul
    Basu, Kanad
    2020 26TH IEEE INTERNATIONAL SYMPOSIUM ON ON-LINE TESTING AND ROBUST SYSTEM DESIGN (IOLTS 2020), 2020,
  • [43] Reconfigurable Multi-Input Adder Design for Deep Neural Network Accelerators
    Moradian, Hossein
    Jo, Sujeong
    Choi, Kiyoung
    2018 INTERNATIONAL SOC DESIGN CONFERENCE (ISOCC), 2018, : 212 - 213
  • [44] Memory Bandwidth and Energy Efficiency Optimization of Deep Convolutional Neural Network Accelerators
    Nie, Zikai
    Li, Zhisheng
    Wang, Lei
    Guo, Shasha
    Dou, Qiang
    ADVANCED COMPUTER ARCHITECTURE, 2018, 908 : 15 - 29
  • [45] Attacking a Joint Protection Scheme for Deep Neural Network Hardware Accelerators and Models
    Wilhelmstaetter, Simon
    Conrad, Joschua
    Upadhyaya, Devanshi
    Polian, Ilia
    Ortmanns, Maurits
    2024 IEEE 6TH INTERNATIONAL CONFERENCE ON AI CIRCUITS AND SYSTEMS, AICAS 2024, 2024, : 144 - 148
  • [46] Efficient Compression Technique for NoC-based Deep Neural Network Accelerators
    Lorandel, Jordane
    Lahdhiri, Habiba
    Bourdel, Emmanuelle
    Monteleone, Salvatore
    Palesi, Maurizio
    2020 23RD EUROMICRO CONFERENCE ON DIGITAL SYSTEM DESIGN (DSD 2020), 2020, : 174 - 179
  • [47] Efficient On-Line Error Detection and Mitigation for Deep Neural Network Accelerators
    Schorn, Christoph
    Guntoro, Andre
    Ascheid, Gerd
    COMPUTER SAFETY, RELIABILITY, AND SECURITY (SAFECOMP 2018), 2018, 11093 : 205 - 219
  • [48] VLSI implementation of transcendental function hyperbolic tangent for deep neural network accelerators
    Rajput, Gunjan
    Raut, Gopal
    Chandra, Mahesh
    Vishvakarma, Santosh Kumar
    MICROPROCESSORS AND MICROSYSTEMS, 2021, 84
  • [49] Understanding Error Propagation in Deep Learning Neural Network (DNN) Accelerators and Applications
    Li, Guanpeng
    Hari, Siva Kumar Sastry
    Sullivan, Michael
    Tsai, Timothy
    Pattabiraman, Karthik
    Emer, Joel
    Keckler, Stephen W.
    SC'17: PROCEEDINGS OF THE INTERNATIONAL CONFERENCE FOR HIGH PERFORMANCE COMPUTING, NETWORKING, STORAGE AND ANALYSIS, 2017,
  • [50] Pico-Programmable Neurons to Reduce Computations for Deep Neural Network Accelerators
    Nahvy, Alireza
    Navabi, Zainalabedin
    IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, 2024, 32 (07) : 1216 - 1227