MEMORY REDUCTION METHOD FOR DEEP NEURAL NETWORK TRAINING

被引:0
|
作者
Shirahata, Koichi [1 ]
Tomita, Yasumoto [1 ]
Ike, Atsushi [1 ]
机构
[1] Fujitsu Labs Ltd, Nakahara Ku, 4-1-1 Kamikodanaka, Kawasaki, Kanagawa 2118588, Japan
关键词
Deep Neural Networks; Memory Management; Accelerators;
D O I
暂无
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Training deep neural networks requires a large amount of memory, making very deep neural networks difficult to fit on accelerator memories. In order to overcome this limitation, we present a method to reduce the amount of memory for training a deep neural network. The method enables to suppress memory increase during the backward pass, by reusing the memory regions allocated for the forward pass. Experimental results exhibit our method reduced the occupied memory size in training by 44.7% on VGGNet with no accuracy affection. Our method also enabled training speedup by increasing the mini batch size up to double.
引用
收藏
页数:6
相关论文
共 50 条
  • [21] Benchmarking and Analyzing Deep Neural Network Training
    Zhu, Hongyu
    Akrout, Mohamed
    Zheng, Bojian
    Pelegris, Andrew
    Jayarajan, Anand
    Phanishayee, Amar
    Schroeder, Bianca
    Pekhimenko, Gennady
    2018 IEEE INTERNATIONAL SYMPOSIUM ON WORKLOAD CHARACTERIZATION (IISWC), 2018, : 88 - 100
  • [22] Deep Neural Network Training with iPSO Algorithm
    Kosten, Mehmet Muzaffer
    Barut, Murat
    Acir, Nurettin
    2018 26TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2018,
  • [23] CLUSTER ADAPTIVE TRAINING FOR DEEP NEURAL NETWORK
    Tan, Tian
    Qian, Yanmin
    Yin, Maofan
    Zhuang, Yimeng
    Yu, Kai
    2015 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING (ICASSP), 2015, : 4325 - 4329
  • [24] PCM: Precision-Controlled Memory System for Energy Efficient Deep Neural Network Training
    Kim, Boyeal
    Lee, Sang Hyun
    Kim, Hyun
    Duy-Thanh Nguyen
    Minh-Son Le
    Chang, Ik Joon
    Kwon, Dohun
    Yoo, Jin Hyeok
    Choi, Jun Won
    Lee, Hyuk-Jae
    PROCEEDINGS OF THE 2020 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE 2020), 2020, : 1199 - 1204
  • [25] The quick method of neural network training
    Dubrovin, V
    Subbotin, S
    MODERN PROBLEMS OF RADIO ENGINEERING, TELECOMMUNICATIONS AND COMPUTER SCIENCE, PROCEEDINGS, 2002, : 266 - 267
  • [26] Phantom-Based Training Framework for Deep Convolutional Neural Network CT Noise Reduction
    Huber, N.
    Missert, A.
    Gong, H.
    Leng, S.
    Yu, L.
    McCollough, C.
    MEDICAL PHYSICS, 2020, 47 (06) : E271 - E271
  • [27] DEEP NEURAL NETWORK (DNN) AUDIO CODER USING A PERCEPTUALLY IMPROVED TRAINING METHOD
    Shin, Seungmin
    Byun, Joon
    Park, Youngcheol
    Sung, Jongmo
    Beack, Seungkwon
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 871 - 875
  • [28] Suppressing seismic multiples based on the deep neural network method with data augmentation training
    Wang KunXi
    Hu TianYue
    Liu XiaoZhou
    Wang ShangXu
    Wei JianXin
    CHINESE JOURNAL OF GEOPHYSICS-CHINESE EDITION, 2021, 64 (11): : 4196 - 4214
  • [29] A training method for deep neural network inference accelerators with high tolerance for their hardware imperfection
    Gao, Shuchao
    Ohsawa, Takashi
    JAPANESE JOURNAL OF APPLIED PHYSICS, 2024, 63 (02)
  • [30] Design of Sports Training Method Based on Multilayer Feature Fusion and Deep Neural Network
    Sun, Xiaojuan
    Saibon, Jamalsafri
    IEEE ACCESS, 2024, 12 : 150204 - 150212