Bit-Beading: Stringing bit-level MAC results for Accelerating Neural Networks

被引:0
|
作者
Anwar, Zeeshan [1 ]
Longchar, Imlijungla [1 ]
Kapoor, Hemangee K. [1 ]
机构
[1] IIT Guwahati, Dept Comp Sci & Engn, Gauhati, India
关键词
MAC Unit; Reconfigurable Arithmetic; Booth's algorithm; CNN; DNN; Neural Network; Low Precision;
D O I
10.1109/VLSID60093.2024.00042
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
On account of the rising demands of AI applications and the consequent need for improvement, researchers are designing better and faster algorithms and architectures. Convolutional Neural Networks (CNN) are neural networks that have become ubiquitous and find applications in the domain of computer vision. Inference in CNN involves convolution operation, which mainly consists of a massive number of matrix multiplications. Optimising these multiplications will enable faster execution of the inference tasks. Fixed precision during inference takes the same time to compute for both higher and lower precision. It is noted in the literature that lowering the precision to some extent does not affect the inference accuracy. In this paper, we propose a reconfigurable multiplier that can handle the precision of different magnitudes. We design Bit-Bead, a basic unit based on Booth's algorithm, where several bit-beads are composed (i.e., stringed) to form a multiplier of the required precision. The reconfigurable multipliers need low latency due to lower precision and also enable performing multiple low-precision computations. Our proposal shows considerable performance improvement compared to the baseline and existing designs.
引用
收藏
页码:216 / 221
页数:6
相关论文
共 50 条
  • [1] Bit Fusion: Bit-Level Dynamically Composable Architecture for Accelerating Deep Neural Networks
    Sharma, Hardik
    Park, Jongse
    Suda, Naveen
    Lai, Liangzhen
    Chau, Benson
    Chandra, Vikas
    Esmaeilzadeh, Hadi
    2018 ACM/IEEE 45TH ANNUAL INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE (ISCA), 2018, : 764 - 775
  • [2] BIT-LEVEL SYNCHRONIZATION IN MICROPROCESSOR NETWORKS
    SINTONEN, L
    UOTILA, P
    IEE PROCEEDINGS-E COMPUTERS AND DIGITAL TECHNIQUES, 1981, 128 (03): : 103 - 106
  • [3] Approach to Improve the Performance Using Bit-level Sparsity in Neural Networks
    Kang, Yesung
    Kwon, Eunji
    Lee, Seunggyu
    Byun, Younghoon
    Lee, Youngjoo
    Kang, Seokhyeong
    PROCEEDINGS OF THE 2021 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE 2021), 2021, : 1516 - 1521
  • [4] Exploiting neural networks bit-level redundancy to mitigate the impact of faults at inference
    Izan Catalán
    José Flich
    Carles Hernández
    The Journal of Supercomputing, 2025, 81 (1)
  • [5] Bit-Balance: Model-Hardware Codesign for Accelerating NNs by Exploiting Bit-Level Sparsity
    Sun, Wenhao
    Zou, Zhiwei
    Liu, Deng
    Sun, Wendi
    Chen, Song
    Kang, Yi
    IEEE TRANSACTIONS ON COMPUTERS, 2024, 73 (01) : 152 - 163
  • [6] Accelerating Fully Homomorphic Encryption by Bridging Modular and Bit-Level Arithmetic
    Chielle, Eduardo
    Mazonka, Oleg
    Gamil, Homer
    Maniatakos, Michail
    2022 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED DESIGN, ICCAD, 2022,
  • [7] Bit-Level Taint Analysis
    Yadegari, Babak
    Debray, Saumya
    2014 14TH IEEE INTERNATIONAL WORKING CONFERENCE ON SOURCE CODE ANALYSIS AND MANIPULATION (SCAM 2014), 2014, : 255 - 264
  • [8] Bit-level Locking for Concurrency Control
    Abbass, Jad F.
    Haraty, Ramzi A.
    2009 IEEE/ACS INTERNATIONAL CONFERENCE ON COMPUTER SYSTEMS AND APPLICATIONS, VOLS 1 AND 2, 2009, : 168 - 173
  • [9] Unconditional bases and bit-level compression
    Donoho, DL
    APPLIED AND COMPUTATIONAL HARMONIC ANALYSIS, 1996, 3 (04) : 388 - 392
  • [10] Bit-level stopping in turbo decoding
    Kim, DH
    Kim, SW
    57TH IEEE VEHICULAR TECHNOLOGY CONFERENCE, VTC 2003-SPRING, VOLS 1-4, PROCEEDINGS, 2003, : 2134 - 2138