Tetris: A Heuristic Static Memory Management Framework for Uniform Memory Multicore Neural Network Accelerators

被引:0
|
作者
Xiao-Bing Chen
Hao Qi
Shao-Hui Peng
Yi-Min Zhuang
Tian Zhi
Yun-Ji Chen
机构
[1] State Key Laboratory of Computer Architecture,
[2] Institute of Computing Technology,undefined
[3] Chinese Academy of Sciences,undefined
[4] University of Chinese Academy of Sciences,undefined
[5] School of Computer Science and Technology,undefined
[6] University of Science and Technology of China,undefined
[7] Chinese Academy of Sciences Center for Excellence in Brain Science and Intelligence Technology,undefined
关键词
multicore neural network accelerator; liveness analysis; static memory management; memory reuse; genetic algorithm;
D O I
暂无
中图分类号
学科分类号
摘要
Uniform memory multicore neural network accelerators (UNNAs) furnish huge computing power to emerging neural network applications. Meanwhile, with neural network architectures going deeper and wider, the limited memory capacity has become a constraint to deploy models on UNNA platforms. Therefore how to efficiently manage memory space and how to reduce workload footprints are urgently significant. In this paper, we propose Tetris: a heuristic static memory management framework for UNNA platforms. Tetris reconstructs execution flows and synchronization relationships among cores to analyze each tensor's liveness interval. Then the memory management problem is converted to a sequence per- mutation problem. Tetris uses a genetic algorithm to explore the permutation space to optimize the memory management strategy and reduce memory footprints. We evaluate several typical neural networks and the experimental results demonstrate that Tetris outperforms the state-of-the-art memory allocation methods, and achieves an average memory reduction ratio of 91.9% and 87.9% for a quad-core and a 16-core Cambricon-X platform, respectively.
引用
收藏
页码:1255 / 1270
页数:15
相关论文
共 50 条
  • [21] GShuttle: Optimizing Memory Access Efficiency for Graph Convolutional Neural Network Accelerators
    Jia-Jun Li
    Ke Wang
    Hao Zheng
    Ahmed Louri
    Journal of Computer Science and Technology, 2023, 38 : 115 - 127
  • [22] Tetris: Re-architecting Convolutional Neural Network Computation for Machine Learning Accelerators
    Lu, Hang
    Wei, Xin
    Lin, Ning
    Yan, Guihai
    Li, Xiao-Wei
    2018 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER-AIDED DESIGN (ICCAD) DIGEST OF TECHNICAL PAPERS, 2018,
  • [23] Analyzing Memory Interference of FPGA Accelerators on Multicore Hosts in Heterogeneous Reconfigurable SoCs
    Mattheeuws, Maxim
    Forsberg, Bjoern
    Kurth, Andreas
    Benini, Luca
    PROCEEDINGS OF THE 2021 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE 2021), 2021, : 1152 - 1155
  • [24] MNEMEE - A Framework for Memory Management and Optimization of Static and Dynamic Data in MPSoCs
    Mallik, Arindam
    Marwedel, Peter
    Soudris, Dimitrios
    Stuijk, Sander
    PROCEEDINGS OF THE 2010 INTERNATIONAL CONFERENCE ON COMPILERS, ARCHITECTURES AND SYNTHESIS FOR EMBEDDED SYSTEMS (CASES '10), 2010, : 257 - 258
  • [25] Scratchpad Memory Management for Deep Learning Accelerators
    Zouzoula, Stavroula
    Maleki, Mohammad Ali
    Azhar, Muhammad Waqar
    Trancoso, Pedro
    53RD INTERNATIONAL CONFERENCE ON PARALLEL PROCESSING, ICPP 2024, 2024, : 629 - 639
  • [26] A memory unit for priority management in IPSec accelerators
    Dadda, Luigi
    Ferrante, Alberto
    Macchetti, Marco
    2007 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS, VOLS 1-14, 2007, : 1533 - +
  • [27] A Method for Reverse Engineering Neural Network Parameters from Compute-in-Memory Accelerators
    Read, James
    Li, Wantong
    Yu, Shimeng
    2022 IEEE COMPUTER SOCIETY ANNUAL SYMPOSIUM ON VLSI (ISVLSI 2022), 2022, : 302 - 307
  • [28] Analog-memory-based In-Memory-Computing Accelerators for Deep Neural Networks
    Tsai, Hsinyu
    2024 IEEE WORKSHOP ON MICROELECTRONICS AND ELECTRON DEVICES, WMED, 2024, : XIII - XIII
  • [29] Memory management technology based on homegrown multicore processor
    Zheng, Yan
    Zou, Tong
    Wang, Xingyan
    CCF TRANSACTIONS ON HIGH PERFORMANCE COMPUTING, 2020, 2 (04) : 376 - 381
  • [30] Memory management technology based on homegrown multicore processor
    Yan Zheng
    Tong Zou
    Xingyan Wang
    CCF Transactions on High Performance Computing, 2020, 2 : 376 - 381