ALPINE: Analog In-Memory Acceleration With Tight Processor Integration for Deep Learning

被引:6
|
作者
Klein, Joshua [1 ]
Boybat, Irem [2 ]
Qureshi, Yasir Mahmood [1 ]
Dazzi, Martino [2 ]
Levisse, Alexandre [1 ]
Ansaloni, Giovanni [1 ]
Zapater, Marina [3 ]
Sebastian, Abu [2 ]
Atienza, David [1 ]
机构
[1] Ecole Polytech Fed Lausanne EPFL, Embedded Syst Lab ESL, CH-1015 Lausanne, Switzerland
[2] BM Res Europe, CH-8803 Ruschlikon, Switzerland
[3] HEIG VD, CH-1401 Yverdon, Switzerland
基金
欧盟地平线“2020”;
关键词
Hardware; Computational modeling; Computer architecture; Biological system modeling; In-memory computing; Reduced instruction set computing; Recurrent neural networks; AI accelerators; architectural exploration; artificial neural networks; gem5; neuromorphic computing; PERFORMANCE;
D O I
10.1109/TC.2022.3230285
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Analog in-memory computing (AIMC) cores offers significant performance and energy benefits for neural network inference with respect to digital logic (e.g., CPUs). AIMCs accelerate matrix-vector multiplications, which dominate these applications' run-time. However, AIMC-centric platforms lack the flexibility of general-purpose systems, as they often have hard-coded data flows and can only support a limited set of processing functions. With the goal of bridging this gap in flexibility, we present a novel system architecture that tightly integrates analog in-memory computing accelerators into multi-core CPUs in general-purpose systems. We developed a powerful gem5-based full system-level simulation framework into the gem5-X simulator, ALPINE, which enables an in-depth characterization of the proposed architecture. ALPINE allows the simulation of the entire computer architecture stack from major hardware components to their interactions with the Linux OS. Within ALPINE, we have defined a custom ISA extension and a software library to facilitate the deployment of inference models. We showcase and analyze a variety of mappings of different neural network types, and demonstrate up to 20.5x/20.8x performance/energy gains with respect to a SIMD-enabled ARM CPU implementation for convolutional neural networks, multi-layer perceptrons, and recurrent neural networks.
引用
收藏
页码:1985 / 1998
页数:14
相关论文
共 50 条
  • [1] Vesti: An In-Memory Computing Processor for Deep Neural Networks Acceleration
    Jiang, Zhewei
    Yin, Shihui
    Kim, Minkyu
    Gupta, Tushar
    Seok, Mingoo
    Seo, Jae-sun
    [J]. CONFERENCE RECORD OF THE 2019 FIFTY-THIRD ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS, 2019, : 1516 - 1521
  • [2] Deep learning acceleration based on in-memory computing
    Eleftheriou, E.
    Le Gallo, M.
    Nandakumar, S. R.
    Piveteau, C.
    Boybat, I
    Joshi, V
    Khaddam-Aljameh, R.
    Dazzi, M.
    Giannopoulos, I
    Karunaratne, G.
    Kersting, B.
    Stanisavljevic, M.
    Jonnalagadda, V. P.
    Ioannou, N.
    Kourtis, K.
    Francese, P. A.
    Sebastian, A.
    [J]. IBM JOURNAL OF RESEARCH AND DEVELOPMENT, 2019, 63 (06)
  • [3] GenPIP: In-Memory Acceleration of Genome Analysis via Tight Integration of Basecalling and Read Mapping
    Mao, Haiyu
    Alser, Mohammed
    Sadrosadati, Mohammad
    Firtina, Can
    Baranwal, Akanksha
    Cali, Damla Senol
    Manglik, Aditya
    Alserr, Nour Almadhoun
    Mutlu, Onur
    [J]. 2022 55TH ANNUAL IEEE/ACM INTERNATIONAL SYMPOSIUM ON MICROARCHITECTURE (MICRO), 2022, : 710 - 726
  • [4] Deep Learning Acceleration using Digital-based Processing In-Memory
    Imani, Mohsen
    Gupta, Saransh
    Kim, Yeseong
    Rosing, Tajana
    [J]. 2020 IEEE 33RD INTERNATIONAL SYSTEM-ON-CHIP CONFERENCE (SOCC), 2020, : 123 - 128
  • [5] In-Memory Computing for Machine Learning and Deep Learning
    Lepri, N.
    Glukhov, A.
    Cattaneo, L.
    Farronato, M.
    Mannocci, P.
    Ielmini, D.
    [J]. IEEE JOURNAL OF THE ELECTRON DEVICES SOCIETY, 2023, 11 : 587 - 601
  • [6] Analog In-Memory Subthreshold Deep Neural Network Accelerator
    Fick, L.
    Blaauw, D.
    Sylvester, D.
    Skrzyniarz, S.
    Parikh, M.
    Fick, D.
    [J]. 2017 IEEE CUSTOM INTEGRATED CIRCUITS CONFERENCE (CICC), 2017,
  • [7] In-memory Learning with Analog Resistive Switching Memory: A Review and Perspective
    Xi, Yue
    Gao, Bin
    Tang, Jianshi
    Chen, An
    Chang, Meng-Fan
    Hu, Xiaobo Sharon
    Spiegel, Jan Van Der
    Qian, He
    Wu, Huaqiang
    [J]. PROCEEDINGS OF THE IEEE, 2021, 109 (01) : 14 - 42
  • [8] Tightly Coupled Machine Learning Coprocessor Architecture With Analog In-Memory Computing for Instruction-Level Acceleration
    Chung, SungWon
    Wang, Jiemi
    [J]. IEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS, 2019, 9 (03) : 544 - 561
  • [9] Optimizing for In-Memory Deep Learning With Emerging Memory Technology
    Wang, Zhehui
    Luo, Tao
    Goh, Rick Siow Mong
    Zhang, Wei
    Wong, Weng-Fai
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, : 1 - 15
  • [10] Towards Resilient Analog In-Memory Deep Learning via Data Layout Re-Organization
    Rashed, Muhammad Rashedul Haq
    Awad, Amro
    Jha, Sumit Kumar
    Ewetz, Rickard
    [J]. PROCEEDINGS OF THE 59TH ACM/IEEE DESIGN AUTOMATION CONFERENCE, DAC 2022, 2022, : 859 - 864