Saving Memory Space in Deep Neural Networks by Recomputing: A Survey

被引:0
|
作者
Ulidowski, Irek [1 ,2 ]
机构
[1] Univ Leicester, Sch Comp & Math Sci, Leicester, Leics, England
[2] AGH Univ Sci & Technol, Dept Appl Informat, Krakow, Poland
来源
关键词
Deep Neural Networks; recomputing activations;
D O I
10.1007/978-3-031-38100-3_7
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Training a multilayered neural network involves execution of the network on the training data, followed by calculating the error between the predicted and actual output, and then performing backpropagation to update the network's weights in order to minimise the overall error. This process is repeated many times, with the network updating its weights until it produces the desired output with a satisfactory level of accuracy. It requires storage in memory of activation and gradient data for each layer during each training run of the network. This paper surveys the main approaches to recomputing the needed activation and gradient data instead of storing it in memory. We discuss how these approaches relate to reversible computation techniques.
引用
收藏
页码:89 / 105
页数:17
相关论文
共 50 条
  • [21] POSTER: Design Space Exploration for Performance Optimization of Deep Neural Networks on Shared Memory Accelerators
    Venkataramani, Swagath
    Choi, Jungwook
    Srinivasan, Vijayalakshmi
    Gopalakrishnan, Kailash
    Chang, Leland
    2017 26TH INTERNATIONAL CONFERENCE ON PARALLEL ARCHITECTURES AND COMPILATION TECHNIQUES (PACT), 2017, : 146 - 147
  • [22] A Survey on Memory Subsystems for Deep Neural Network Accelerators
    Asad, Arghavan
    Kaur, Rupinder
    Mohammadi, Farah
    FUTURE INTERNET, 2022, 14 (05):
  • [23] Accelerating Deep Neural Networks with Analog Memory Devices
    Ambrogio, Stefano
    Narayanan, Pritish
    Tsai, Hsinyu
    Mackin, Charles
    Spoon, Katherine
    Chen, An
    Fasoli, Andrea
    Friz, Alexander
    Burr, Geoffrey W.
    2020 2ND IEEE INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE CIRCUITS AND SYSTEMS (AICAS 2020), 2020, : 149 - 152
  • [24] ACCELERATING DEEP NEURAL NETWORKS WITH ANALOG MEMORY DEVICES
    Burr, Geoffrey W.
    Ambrogio, Stefano
    Narayanan, Pritish
    Tsai, Hsinyu
    Mackin, Charles
    Chen, An
    2019 CHINA SEMICONDUCTOR TECHNOLOGY INTERNATIONAL CONFERENCE (CSTIC), 2019,
  • [25] Accelerating Deep Neural Networks with Analog Memory Devices
    Spoon, Katie
    Ambrogio, Stefano
    Narayanan, Pritish
    Tsai, Hsinyu
    Mackin, Charles
    Chen, An
    Fasoli, Andrea
    Friz, Alexander
    Burr, Geoffrey W.
    2020 IEEE INTERNATIONAL MEMORY WORKSHOP (IMW 2020), 2020, : 111 - 114
  • [26] Practical Attacks on Deep Neural Networks by Memory Trojaning
    Hu, Xing
    Zhao, Yang
    Deng, Lei
    Liang, Ling
    Zuo, Pengfei
    Ye, Jing
    Lin, Yingyan
    Xie, Yuan
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2021, 40 (06) : 1230 - 1243
  • [27] Video Summarization Using Deep Neural Networks: A Survey
    Apostolidis, Evlampios
    Adamantidou, Eleni
    Metsai, Alexandros, I
    Mezaris, Vasileios
    Patras, Ioannis
    PROCEEDINGS OF THE IEEE, 2021, 109 (11) : 1838 - 1863
  • [28] Structured Pruning for Deep Convolutional Neural Networks: A Survey
    He, Yang
    Xiao, Lingao
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (05) : 2900 - 2919
  • [29] A survey of the recent architectures of deep convolutional neural networks
    Asifullah Khan
    Anabia Sohail
    Umme Zahoora
    Aqsa Saeed Qureshi
    Artificial Intelligence Review, 2020, 53 : 5455 - 5516
  • [30] Topic Modelling Meets Deep Neural Networks: A Survey
    Zhao, He
    Dinh Phung
    Viet Huynh
    Jin, Yuan
    Du, Lan
    Buntine, Wray
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 4713 - 4720