An In-Memory Analog Computing Co-Processor for Energy-Efficient CNN Inference on Mobile Devices

被引:11
|
作者
Elbtity, Mohammed [1 ]
Singh, Abhishek [2 ]
Reidy, Brendan [1 ]
Guo, Xiaochen [2 ]
Zand, Ramtin [1 ]
机构
[1] Univ South Carolina, Dept Comp Sci & Engn, Columbia, SC 29208 USA
[2] Lehigh Univ, Dept Elect & Comp Engn, Bethlehem, PA 18015 USA
关键词
in-memory computing; magnetic random access memory (MRAM); convolutional neural networks (CNNs); mixed-precision and mixed-signal inference;
D O I
10.1109/ISVLSI51109.2021.00043
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we develop an in-memory analog computing (IMAC) architecture realizing both synaptic behavior and activation functions within non-volatile memory arrays. Spin-orbit torque magnetoresistive random-access memory (SOT-MRAM) devices are leveraged to realize sigmoidal neurons as well as binarized synapses. First, it is shown the proposed IMAC architecture can be utilized to realize a multilayer perceptron (MLP) classifier achieving orders of magnitude performance improvement compared to previous mixed-signal and digital implementations. Next, a heterogeneous mixed-signal and mixed-precision CPU-IMAC architecture is proposed for convolutional neural networks (CNNs) inference on mobile processors, in which IMAC is designed as a co-processor to realize fully-connected (FC) layers whereas convolution layers are executed in CPU. Architecture-level analytical models are developed to evaluate the performance and energy consumption of the CPU-IMAC architecture. Simulation results exhibit 6.5% and 10% energy savings for CPU-IMAC based realizations of LeNet and VGG CNN models, for MNIST and CIFAR-10 pattern recognition tasks, respectively.
引用
收藏
页码:188 / 193
页数:6
相关论文
共 50 条
  • [41] AnalogNAS: A Neural Network Design Framework for Accurate Inference with Analog In-Memory Computing
    Benmeziane, Hadjer
    Lammie, Corey
    Boybat, Irem
    Rasch, Malte
    Le Gallo, Manuel
    Tsai, Hsinyu
    Muralidhar, Ramachandran
    Niar, Smail
    Hamza, Ouarnoughi
    Narayanan, Vijay
    Sebastian, Abu
    El Maghraoui, Kaoutar
    2023 IEEE INTERNATIONAL CONFERENCE ON EDGE COMPUTING AND COMMUNICATIONS, EDGE, 2023, : 233 - 244
  • [42] Energy-Efficient Computation Offloading for Wearable Devices and Smartphones in Mobile Cloud Computing
    Ragona, Claudio
    Granelli, Fabrizio
    Fiandrino, Claudio
    Kliazovich, Dzmitry
    Bouvry, Pascal
    2015 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2015,
  • [43] An Energy-Efficient In-Memory Accelerator for Graph Construction and Updating
    Chen, Mingkai
    Liu, Cheng
    Liang, Shengwen
    He, Lei
    Wang, Ying
    Zhang, Lei
    Li, Huawei
    Li, Xiaowei
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2024, 43 (06) : 1781 - 1793
  • [44] Evaluating the Vector Supercomputer SX-Aurora TSUBASA as a Co-Processor for In-Memory Database Systems
    Johannes Pietrzyk
    Dirk Habich
    Patrick Damme
    Erich Focht
    Wolfgang Lehner
    Datenbank-Spektrum, 2019, 19 (3) : 183 - 197
  • [45] OCP: Offload Co-Processor for Energy Efficiency in Embedded Mobile Systems
    Tang, Jie
    Liu, Chen
    Chou, Yu-Liang
    Liu, Shaoshan
    PROCEEDINGS OF THE 2013 IEEE 24TH INTERNATIONAL CONFERENCE ON APPLICATION-SPECIFIC SYSTEMS, ARCHITECTURES AND PROCESSORS (ASAP 13), 2013, : 107 - 110
  • [46] Design of High-RA STT-MRAM for Future Energy-Efficient In-Memory Computing
    Hong, Ming-Chun
    Su, Yi-Hui
    Chen, Guan-Long
    Hsin, Yu-Chen
    Chang, Yao-Jen
    Chen, Kuan-Ming
    Yang, Shan-Yi
    Wang, I-Jung
    Rahaman, S. K. Ziaur
    Lee, Hsin-Han
    Wei, Jeng-Hua
    Sheu, Shyh-Shyuan
    Lo, Wei-Chung
    Chang, Shih-Chieh
    Hou, Tuo-Hung
    2023 INTERNATIONAL VLSI SYMPOSIUM ON TECHNOLOGY, SYSTEMS AND APPLICATIONS, VLSI-TSA/VLSI-DAT, 2023,
  • [47] Reconfigurable logic in nanosecond Cu/GeTe/TiN filamentary memristors for energy-efficient in-memory computing
    Jin, Miao-Miao
    Cheng, Long
    Li, Yi
    Hu, Si-Yu
    Lu, Ke
    Chen, Jia
    Duan, Nian
    Wang, Zhuo-Rui
    Zhou, Ya-Xiong
    Chang, Ting-Chang
    Miao, Xiang-Shui
    NANOTECHNOLOGY, 2018, 29 (38)
  • [48] An energy-efficient in-memory computing architecture for survival data analysis based on resistive switching memories
    Baroni, Andrea
    Glukhov, Artem
    Perez, Eduardo
    Wenger, Christian
    Calore, Enrico
    Schifano, Sebastiano Fabio
    Olivo, Piero
    Ielmini, Daniele
    Zambelli, Cristian
    FRONTIERS IN NEUROSCIENCE, 2022, 16
  • [49] Reconfigurable logic in nanosecond Cu/GeTe/TiN filamentary memristors for energy-efficient in-memory computing
    Jin, Miao-Miao
    Cheng, Long
    Li, Yi
    Hu, Si-Yu
    Lu, Ke
    Chen, Jia
    Duan, Nian
    Wang, Zhuo-Rui
    Zhou, Ya-Xiong
    Chang, Ting-Chang
    Miao, Xiang-Shui
    Nanotechnology, 2018, 29 (38):
  • [50] The Opportunity of Negative Capacitance Behavior in Flash Memory for High-Density and Energy-Efficient In-Memory Computing Applications
    Kim, Taeho
    Kim, Giuk
    Lee, Young Kyu
    Ko, Dong Han
    Hwang, Junghyeon
    Lee, Sangho
    Shin, Hunbeom
    Jeong, Yeongseok
    Jung, Seong-Ook
    Jeon, Sanghun
    ADVANCED FUNCTIONAL MATERIALS, 2023, 33 (07)