Tetris: A Heuristic Static Memory Management Framework for Uniform Memory Multicore Neural Network Accelerators

被引:0
|
作者
Xiao-Bing Chen
Hao Qi
Shao-Hui Peng
Yi-Min Zhuang
Tian Zhi
Yun-Ji Chen
机构
[1] State Key Laboratory of Computer Architecture,
[2] Institute of Computing Technology,undefined
[3] Chinese Academy of Sciences,undefined
[4] University of Chinese Academy of Sciences,undefined
[5] School of Computer Science and Technology,undefined
[6] University of Science and Technology of China,undefined
[7] Chinese Academy of Sciences Center for Excellence in Brain Science and Intelligence Technology,undefined
关键词
multicore neural network accelerator; liveness analysis; static memory management; memory reuse; genetic algorithm;
D O I
暂无
中图分类号
学科分类号
摘要
Uniform memory multicore neural network accelerators (UNNAs) furnish huge computing power to emerging neural network applications. Meanwhile, with neural network architectures going deeper and wider, the limited memory capacity has become a constraint to deploy models on UNNA platforms. Therefore how to efficiently manage memory space and how to reduce workload footprints are urgently significant. In this paper, we propose Tetris: a heuristic static memory management framework for UNNA platforms. Tetris reconstructs execution flows and synchronization relationships among cores to analyze each tensor's liveness interval. Then the memory management problem is converted to a sequence per- mutation problem. Tetris uses a genetic algorithm to explore the permutation space to optimize the memory management strategy and reduce memory footprints. We evaluate several typical neural networks and the experimental results demonstrate that Tetris outperforms the state-of-the-art memory allocation methods, and achieves an average memory reduction ratio of 91.9% and 87.9% for a quad-core and a 16-core Cambricon-X platform, respectively.
引用
收藏
页码:1255 / 1270
页数:15
相关论文
共 50 条
  • [41] Memory level neural network: A time-varying neural network for memory input processing
    Gong, Chao
    Zhou, Xianwei
    Lu, Xing
    Lin, Fuhong
    NEUROCOMPUTING, 2021, 425 : 256 - 265
  • [42] In-Network Memory Access Ordering for Heterogeneous Multicore Systems
    Yin, Jieming
    Zhai, Antonia
    2020 14TH IEEE/ACM INTERNATIONAL SYMPOSIUM ON NETWORKS-ON-CHIP (NOCS), 2020,
  • [43] PIMulator-NN: An Event-Driven, Cross-Level Simulation Framework for Processing-In-Memory-Based Neural Network Accelerators
    Zheng, Qilin
    Li, Xingchen
    Guan, Yijin
    Wang, Zongwei
    Cai, Yimao
    Chen, Yiran
    Sun, Guangyu
    Huang, Ru
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2022, 41 (12) : 5464 - 5475
  • [44] Management and Optimization for Nonvolatile Memory-Based Hybrid Scratchpad Memory on Multicore Embedded Processors
    Hu, Jingtong
    Zhuge, Qingfeng
    Xue, Chun Jason
    Tseng, Wei-Che
    Sha, Edwin H. -M.
    ACM TRANSACTIONS ON EMBEDDED COMPUTING SYSTEMS, 2014, 13 (04)
  • [45] Reflections on memory of work and organizational memory in the framework of knowledge management
    Rossetti, Adroaldo G.
    Viegas, Claudia V.
    Vidal, Cristiane Denise
    Costa, Eliete Oliveira
    Fialho, Francisco A. P.
    Paulista dos Santos, Geralda P.
    Meneguelle, Flaviana Cristina
    Goulart Braga, Marta C.
    Sartori, Rejane
    da Silva Filho, Vasco Pinto
    Goncalves Werutsky, Viviane D'Barsoles
    GESTAO E DESENVOLVIMENTO, 2005, 2 (02): : 15 - 20
  • [46] Shared memory resources allocation and management research on multicore systems
    Gao, Ke
    Chen, Li-Cheng
    Fan, Dong-Rui
    Liu, Zhi-Yong
    Jisuanji Xuebao/Chinese Journal of Computers, 2015, 38 (05): : 1020 - 1034
  • [47] POMMEL: Exploring Off-Chip Memory Energy & Power Consumption in Convolutional Neural Network Accelerators
    Montgomerie-Corcoran, Alexander
    Bouganis, Christos-Savvas
    2021 24TH EUROMICRO CONFERENCE ON DIGITAL SYSTEM DESIGN (DSD 2021), 2021, : 442 - 448
  • [48] Memory Recall: A Simple Neural Network Training Framework Against Catastrophic Forgetting
    Zhang, Baosheng
    Guo, Yuchen
    Li, Yipeng
    He, Yuwei
    Wang, Haoqian
    Dai, Qionghai
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (05) : 2010 - 2022
  • [49] A Memory-Efficient Hybrid Parallel Framework for Deep Neural Network Training
    Li, Dongsheng
    Li, Shengwei
    Lai, Zhiquan
    Fu, Yongquan
    Ye, Xiangyu
    Cai, Lei
    Qiao, Linbo
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2024, 35 (04) : 577 - 591
  • [50] Study on Neural Network Integration Method Based on Morphological Associative Memory Framework
    Naiqin Feng
    Xiuqin Geng
    Bin Sun
    Neural Processing Letters, 2021, 53 : 3915 - 3945