High-Throughput In-Memory Computing for Binary Deep Neural Networks With Monolithically Integrated RRAM and 90-nm CMOS

被引:106
|
作者
Yin, Shihui [1 ]
Sun, Xiaoyu [2 ]
Yu, Shimeng [2 ]
Seo, Jae-Sun [1 ]
机构
[1] Arizona State Univ, Sch Elect Comp & Energy Engn, Tempe, AZ 85287 USA
[2] Georgia Inst Technol, Sch Elect & Comp Engn, Atlanta, GA 30332 USA
基金
美国国家科学基金会;
关键词
Quantization (signal); Resistance; Random access memory; Hardware; Decoding; Parallel processing; Prototypes; Deep neural networks (DNNs); in-memory computing (IMC); monolithic integration; nonvolatile memory (NVM); resistive RAM (RRAM);
D O I
10.1109/TED.2020.3015178
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Deep neural network (DNN) hardware designs have been bottlenecked by conventional memories, such as SRAM due to density, leakage, and parallel computing challenges. Resistive devices can address the density and volatility issues but have been limited by peripheral circuit integration. In this work, we present a resistive RAM (RRAM)-based in-memory computing (IMC) design, which is fabricated in 90-nm CMOS with monolithic integration of RRAM devices. We integrated a 128 x 64 RRAM array with CMOS peripheral circuits, including row/column decoders and flash analog-to-digital converters (ADCs), which collectively become a core component for scalable RRAM-based IMC for large DNNs. To maximize IMC parallelism, we assert all 128 wordlines of the RRAM array simultaneously, perform analog computing along the bitlines, and digitize the bitline voltages using ADCs. The resistance distribution of low-resistance states is tightened by an iterative write-verify scheme. Prototype chip measurements demonstrate high binary DNN accuracy of 98.5% for MNIST and 83.5% for CIFAR-10 data sets, with 24 TOPS/W and 158 GOPS. This represents 22.3x and 10.1x improvements in throughput and energy-delay product (EDP), respectively, compared with the state-of-the-art literature, which can enable intelligent functionalities for area-/energy-constrained edge computing devices.
引用
收藏
页码:4185 / 4192
页数:8
相关论文
共 50 条
  • [1] Monolithically Integrated RRAM- and CMOS-Based In-Memory Computing Optimizations for Efficient Deep Learning
    Yin, Shihui
    Kim, Yulhwa
    Han, Xu
    Barnaby, Hugh
    Yu, Shimeng
    Luo, Yandong
    He, Wangxin
    Sun, Xiaoyu
    Kim, Jae-Joon
    Seo, Jae-sun
    [J]. IEEE MICRO, 2019, 39 (06) : 54 - 63
  • [2] RRAM-Based In-Memory Computing for Embedded Deep Neural Networks
    Bankman, D.
    Messner, J.
    Gural, A.
    Murmann, B.
    [J]. CONFERENCE RECORD OF THE 2019 FIFTY-THIRD ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS, 2019, : 1511 - 1515
  • [3] Structured Pruning of RRAM Crossbars for Efficient In-Memory Computing Acceleration of Deep Neural Networks
    Meng, Jian
    Yang, Li
    Peng, Xiaochen
    Yu, Shimeng
    Fan, Deliang
    Seo, Jae-Sun
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2021, 68 (05) : 1576 - 1580
  • [4] Dual-configuration in-memory computing bitcells using SiOx RRAM for binary neural networks
    Kingra, Sandeep Kaur
    Parmar, Vivek
    Negi, Shubham
    Bricalli, Alessandro
    Piccolboni, Giuseppe
    Regev, Amir
    Nodin, Jean-Francois
    Molas, Gabriel
    Suri, Manan
    [J]. APPLIED PHYSICS LETTERS, 2022, 120 (03)
  • [5] A Low-Power High-Throughput In-Memory CMOS-ReRAM Accelerator for Large-Scale Deep Residual Neural Networks
    Cheng, Yuan
    Wong, Ngai
    Liu, Xiong
    Ni, Leibin
    Chen, Hai-Bao
    Yu, Hao
    [J]. 2019 IEEE 13TH INTERNATIONAL CONFERENCE ON ASIC (ASICON), 2019,
  • [6] MOL-Based In-Memory Computing of Binary Neural Networks
    Ali, Khaled Alhaj
    Baghdadi, Amer
    Dupraz, Elsa
    Leonardon, Mathieu
    Rizk, Mostafa
    Diguet, Jean-Philippe
    [J]. IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, 2022, 30 (07) : 869 - 880
  • [7] XNOR-SRAM: In-Memory Computing SRAM Macro for Binary/Ternary Deep Neural Networks
    Jiang, Zhewei
    Yin, Shihui
    Seok, Mingoo
    Seo, Jae-sun
    [J]. 2018 IEEE SYMPOSIUM ON VLSI TECHNOLOGY, 2018, : 173 - 174
  • [8] XNOR-SRAM: In-Memory Computing SRAM Macro for Binary/Ternary Deep Neural Networks
    Yin, Shihui
    Jiang, Zhewei
    Seo, Jae-Sun
    Seok, Mingoo
    [J]. IEEE JOURNAL OF SOLID-STATE CIRCUITS, 2020, 55 (06) : 1733 - 1743
  • [9] Hybrid In-memory Computing Architecture for the Training of Deep Neural Networks
    Joshi, Vinay
    He, Wangxin
    Seo, Jae-sun
    Rajendran, Bipin
    [J]. 2021 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2021,
  • [10] Vesti: An In-Memory Computing Processor for Deep Neural Networks Acceleration
    Jiang, Zhewei
    Yin, Shihui
    Kim, Minkyu
    Gupta, Tushar
    Seok, Mingoo
    Seo, Jae-sun
    [J]. CONFERENCE RECORD OF THE 2019 FIFTY-THIRD ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS, 2019, : 1516 - 1521