Eidetic: An In-Memory Matrix Multiplication Accelerator for Neural Networks

被引:3
|
作者
Eckert, Charles [1 ]
Subramaniyan, Arun [1 ]
Wang, Xiaowei [1 ]
Augustine, Charles [2 ]
Iyer, Ravishankar [3 ]
Das, Reetuparna [1 ]
机构
[1] Univ Michigan, Dept Comp Sci & Engn, Ann Arbor, MI 48109 USA
[2] Intel Corp, Circuit Res Labs, Hillsboro, OR 97124 USA
[3] Intel, Syst Technol Lab, Hillsboro, OR 97124 USA
关键词
B.6.1.e memory used as logic; C.1.3.i neural nets accelerator; C.1.3.e dataflow architectures; MACRO;
D O I
10.1109/TC.2022.3214151
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
This paper presents the Eidetic architecture, which is an SRAM-based ASIC neural network accelerator that eliminates the need to continuously load weights from off-chip, while also minimizing the need to go off chip for intermediate results. Using in-situ arithmetic in the SRAM arrays, this architecture can supports a variety of precision types allowing for effective inference. We also present different data mapping policies for matrix-vector based networks (RNN and MLP) on the Eidetic architecture and describe the tradeoffs involved. With this architecture, multiple layers of a network can be concurrently mapped, storing both the layer weights and intermediate results on-chip, removing the energy and latency penalty of off-chip memory accesses. We evaluate Eidetic on Google's Neural Machine Translation System (GNMT) encoder and demonstrate a 17.20x increase in throughput and 7.77x reduction in average latency over a single TPUv2 chip.
引用
收藏
页码:1539 / 1553
页数:15
相关论文
共 50 条
  • [1] TFix: Exploiting the Natural Redundancy of Ternary Neural Networks for Fault Tolerant In-Memory Vector Matrix Multiplication
    Malhotra, Akul
    Wang, Chunguang
    Gupta, Sumeet Kumar
    [J]. 2023 60TH ACM/IEEE DESIGN AUTOMATION CONFERENCE, DAC, 2023,
  • [2] FAT: An In-Memory Accelerator With Fast Addition for Ternary Weight Neural Networks
    Zhu, Shien
    Duong, Luan H. K.
    Chen, Hui
    Liu, Di
    Liu, Weichen
    [J]. IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2023, 42 (03) : 781 - 794
  • [3] TiM-DNN: Ternary In-Memory Accelerator for Deep Neural Networks
    Jain, Shubham
    Gupta, Sumeet Kumar
    Raghunathan, Anand
    [J]. IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, 2020, 28 (07) : 1567 - 1577
  • [4] In-Memory Computing Based Hardware Accelerator Module for Deep Neural Networks
    Appukuttan, Allen
    Thomas, Emmanuel
    Nair, Harinandan R.
    Hemanth, S.
    Dhanaraj, K. J.
    Azeez, Maleeha Abdul
    [J]. 2022 IEEE 19TH INDIA COUNCIL INTERNATIONAL CONFERENCE, INDICON, 2022,
  • [5] An Energy-efficient Matrix Multiplication Accelerator by Distributed In-memory Computing on Binary RRAM Crossbar
    Ni, Leibin
    Wang, Yuhao
    Yu, Hao
    Yang, Wei
    Weng, Chuliang
    Zhao, Junfeng
    [J]. 2016 21ST ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC), 2016, : 280 - 285
  • [6] An Efficient Optical Sparse Matrix Multiplication Accelerator for Graph Neural Networks
    Jia, Ying
    Guo, Hongxiang
    Guo, Yi
    Wu, Jian
    [J]. 2022 ASIA COMMUNICATIONS AND PHOTONICS CONFERENCE, ACP, 2022, : 1868 - 1872
  • [7] In-memory Photonic Tensor Core Accelerator for Neural Networks-based Applications
    Meng, Jiawei
    Ma, Xiaoxuan
    Peserico, Nicola
    Dalir, Hamed
    Sorger, Volker J.
    [J]. 2023 IEEE PHOTONICS SOCIETY SUMMER TOPICALS MEETING SERIES, SUM, 2023,
  • [8] Vesti: Energy-Efficient In-Memory Computing Accelerator for Deep Neural Networks
    Yin, Shihui
    Jiang, Zhewei
    Kim, Minkyu
    Gupta, Tushar
    Seok, Mingoo
    Seo, Jae-Sun
    [J]. IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, 2020, 28 (01) : 48 - 61
  • [9] Rapid In-Memory Matrix Multiplication Using Associative Processor
    Neggaz, Mohamed Ayoub
    Yantir, Hasan Erdem
    Niar, Smail
    Eltawil, Ahmed
    Kurdahi, Fadi
    [J]. PROCEEDINGS OF THE 2018 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE), 2018, : 985 - 990
  • [10] Scalable In-Memory Computing Architectures for Sparse Matrix Multiplication
    Kendall, Jack D.
    Conklin, Alexander A.
    Pantone, Ross
    Nino, Juan C.
    Kumar, Suhas
    [J]. 2022 INTERNATIONAL ELECTRON DEVICES MEETING, IEDM, 2022,