An Energy-Efficient Time Domain Based Compute In-Memory Architecture for Binary Neural Network

被引:0
|
作者
Chakraborty, Subhradip [1 ]
Kushwaha, Dinesh [2 ]
Goel, Abhishek [2 ]
Singla, Anmol [3 ]
Bulusu, Anand [2 ]
Dasgupta, Sudeb [2 ]
机构
[1] RGIPT Uttar Pradesh, EE Engn Dept, Jais, India
[2] Indian Inst Technol, Elect & Commun Engn Dept, Roorkee, Uttar Pradesh, India
[3] NIT Uttarakhand, ECE Dept, Srinagar, India
关键词
Analog; binary neural network (BNN); compute in-memory (CIM); energy efficiency; time domain computing;
D O I
10.1109/ISQED60706.2024.10528729
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
This paper presents an energy-efficient time domain-based compute in-memory architecture to accelerate the deep neural networks (DNNs). This work focuses on improving the energy efficiency of the multiplication and accumulation (MAC) operation by performing it within the memory cell itself. The proposed approach utilizes time domain computing, which involves introducing a specific delay to a reference signal to perform MAC operations. To convert the time domain signal into a digital form, a time-to-digital converter (TDC) is employed. A 12T time domain-based bit cell generates the necessary delay, while a flash type TDC is used for time to digital conversion. The designed architecture has been implemented using a 45 nm CMOS technology, resulting in the development of a 128x64 SRAM CIM macro. Simulation results demonstrate that the proposed architecture achieved an energy efficiency of 941 TOPS/W at a frequency of 0.5 MHz and 1 V, which is 1.75x higher than the state-of-the-art. Furthermore, the system attained an inference accuracy of 96.7% and 84.52% when tested on the MNIST and CIFAR- 10 datasets, respectively.
引用
收藏
页数:6
相关论文
共 50 条
  • [41] An Energy-Efficient In-Memory Accelerator for Graph Construction and Updating
    Chen, Mingkai
    Liu, Cheng
    Liang, Shengwen
    He, Lei
    Wang, Ying
    Zhang, Lei
    Li, Huawei
    Li, Xiaowei
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2024, 43 (06) : 1781 - 1793
  • [42] iMAT: Energy-Efficient In-Memory Acceleration for Ternary Neural Networks With Sparse Dot Product
    Zhu, Shien
    Huai, Shuo
    Xiong, Guochu
    Liu, Weichen
    2023 IEEE/ACM INTERNATIONAL SYMPOSIUM ON LOW POWER ELECTRONICS AND DESIGN, ISLPED, 2023,
  • [43] Energy-Efficient Associative Memory Based on Neural Cliques
    Boguslawski, Bartosz
    Heitzmann, Frederic
    Larras, Benoit
    Seguin, Fabrice
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2016, 63 (04) : 376 - 380
  • [44] Optimization of DRAM based PIM Architecture for Energy-Efficient Deep Neural Network Training
    Sudarshan, Chirag
    Sadi, Mohammad Hassani
    Weis, Christian
    Wehn, Norbert
    2022 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS 22), 2022, : 1472 - 1476
  • [45] An Energy-Efficient Graphene-based Spiking Neural Network Architecture for Pattern Recognition
    Laurenciu, Nicoleta Cucu
    Timmermans, Charles
    Cotofana, Sorin D.
    2024 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS, ISCAS 2024, 2024,
  • [46] NNPIM: A Processing In-Memory Architecture for Neural Network Acceleration
    Gupta, Saransh
    Imani, Mohsen
    Kaur, Harveen
    Rosing, Tajana Simunic
    IEEE TRANSACTIONS ON COMPUTERS, 2019, 68 (09) : 1325 - 1337
  • [47] NUTS-BSNN: A non-uniform time-step binarized spiking neural network with energy-efficient in-memory computing macro
    Dinh, Van-Ngoc
    Bui, Ngoc-My
    Nguyen, Van-Tinh
    John, Deepu
    Lin, Long-Yang
    Trinh, Quang-Kien
    NEUROCOMPUTING, 2023, 560
  • [48] Sandwich-RAM: An Energy-Efficient In-Memory BWN Architecture with Pulse-Width Modulation
    Yang, Jun
    Kong, Yuyao
    Wang, Zhen
    Liu, Yan
    Wang, Bo
    Yin, Shouyi
    Shi, Longxin
    2019 IEEE INTERNATIONAL SOLID-STATE CIRCUITS CONFERENCE (ISSCC), 2019, 62 : 394 - +
  • [49] Hybrid Convolution Architecture for Energy-Efficient Deep Neural Network Processing
    Kim, Suchang
    Jo, Jihyuck
    Park, In-Cheol
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2021, 68 (05) : 2017 - 2029
  • [50] SparseMEM: Energy-efficient Design for In-memory Sparse-based Graph Processing
    Zahedi, Mahdi
    Custers, Geert
    Shahroodi, Taha
    Gaydadjiev, Georgi
    Wong, Stephan
    Hamdioui, Said
    2023 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION, DATE, 2023,