Deep learning acceleration based on in-memory computing

被引:23
|
作者
Eleftheriou, E. [1 ]
Le Gallo, M. [1 ]
Nandakumar, S. R. [1 ]
Piveteau, C. [1 ]
Boybat, I [1 ]
Joshi, V [1 ]
Khaddam-Aljameh, R. [1 ]
Dazzi, M. [1 ]
Giannopoulos, I [1 ]
Karunaratne, G. [1 ]
Kersting, B. [1 ]
Stanisavljevic, M. [1 ]
Jonnalagadda, V. P. [1 ]
Ioannou, N. [1 ]
Kourtis, K. [1 ]
Francese, P. A. [1 ]
Sebastian, A. [1 ]
机构
[1] IBM Res Zurich, CH-8803 Ruschlikon, Switzerland
关键词
PHASE-CHANGE MEMORY; NETWORK;
D O I
10.1147/JRD.2019.2947008
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Performing computations on conventional von Neumann computing systems results in a significant amount of data being moved back and forth between the physically separated memory and processing units. This costs time and energy, and constitutes an inherent performance bottleneck. In-memory computing is a novel non-von Neumann approach, where certain computational tasks are performed in the memory itself. This is enabled by the physical attributes and state dynamics of memory devices, in particular, resistance-based nonvolatile memory technology. Several computational tasks such as logical operations, arithmetic operations, and even certain machine learning tasks can be implemented in such a computational memory unit. In this article, we first introduce the general notion of in-memory computing and then focus on mixed-precision deep learning training with in-memory computing. The efficacy of this new approach will be demonstrated by training the MNIST multilayer perceptron network achieving high accuracy. Moreover, we show how the precision of in-memory computing can be further improved through architectural and device-level innovations. Finally, we present system aspects, such as high-level system architecture, including core-to-core interconnect technologies, and high-level ideas and concepts of the software stack.
引用
收藏
页数:18
相关论文
共 50 条
  • [1] In-Memory Computing for Machine Learning and Deep Learning
    Lepri, N.
    Glukhov, A.
    Cattaneo, L.
    Farronato, M.
    Mannocci, P.
    Ielmini, D.
    [J]. IEEE JOURNAL OF THE ELECTRON DEVICES SOCIETY, 2023, 11 : 587 - 601
  • [2] AI: From Deep Learning to In-Memory Computing
    Lung, Hsiang-Lan
    [J]. METROLOGY, INSPECTION, AND PROCESS CONTROL FOR MICROLITHOGRAPHY XXXIII, 2019, 10959
  • [3] Deep Learning Acceleration using Digital-based Processing In-Memory
    Imani, Mohsen
    Gupta, Saransh
    Kim, Yeseong
    Rosing, Tajana
    [J]. 2020 IEEE 33RD INTERNATIONAL SYSTEM-ON-CHIP CONFERENCE (SOCC), 2020, : 123 - 128
  • [4] Vesti: An In-Memory Computing Processor for Deep Neural Networks Acceleration
    Jiang, Zhewei
    Yin, Shihui
    Kim, Minkyu
    Gupta, Tushar
    Seok, Mingoo
    Seo, Jae-sun
    [J]. CONFERENCE RECORD OF THE 2019 FIFTY-THIRD ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS, 2019, : 1516 - 1521
  • [5] Acceleration of HadoopMapReduce using in-memory Computing
    Seelam, Siva Kumar
    Pattabiraman, V
    [J]. PROCEEDINGS OF THE 2018 INTERNATIONAL CONFERENCE ON RECENT TRENDS IN ADVANCED COMPUTING (ICRTAC-CPS 2018), 2018, : 91 - 96
  • [6] Digital In-Memory Computing to Accelerate Deep Learning Inference on the Edge
    Perri, Stefania
    Zambelli, Cristian
    Ielmini, Daniele
    Silvano, Cristina
    [J]. 2024 IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM WORKSHOPS, IPDPSW 2024, 2024, : 130 - 133
  • [7] ALPINE: Analog In-Memory Acceleration With Tight Processor Integration for Deep Learning
    Klein, Joshua
    Boybat, Irem
    Qureshi, Yasir Mahmood
    Dazzi, Martino
    Levisse, Alexandre
    Ansaloni, Giovanni
    Zapater, Marina
    Sebastian, Abu
    Atienza, David
    [J]. IEEE TRANSACTIONS ON COMPUTERS, 2023, 72 (07) : 1985 - 1998
  • [8] Memristors-From In-Memory Computing, Deep Learning Acceleration, and Spiking Neural Networks to the Future of Neuromorphic and Bio-Inspired Computing
    Mehonic, Adnan
    Abu Sebastian
    Rajendran, Bipin
    Simeone, Osvaldo
    Vasilaki, Eleni
    Kenyon, Anthony J.
    [J]. ADVANCED INTELLIGENT SYSTEMS, 2020, 2 (11)
  • [9] Structured Pruning of RRAM Crossbars for Efficient In-Memory Computing Acceleration of Deep Neural Networks
    Meng, Jian
    Yang, Li
    Peng, Xiaochen
    Yu, Shimeng
    Fan, Deliang
    Seo, Jae-Sun
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2021, 68 (05) : 1576 - 1580
  • [10] Deep learning incorporating biologically inspired neural dynamics and in-memory computing
    Stanisław Woźniak
    Angeliki Pantazi
    Thomas Bohnstingl
    Evangelos Eleftheriou
    [J]. Nature Machine Intelligence, 2020, 2 : 325 - 336