Low-Rank Gradient Descent for Memory-Efficient Training of Deep In-Memory Arrays

被引:0
|
作者
Huang, Siyuan [1 ]
Hoskins, Brian D. [2 ]
Daniels, Matthew W. [2 ]
Stiles, Mark D. [2 ]
Adam, Gina C. [3 ]
机构
[1] George Washington Univ, Dept Comp Sci, Washington, DC 20038 USA
[2] Natl Inst Stand & Technol, Gaithersburg, MD USA
[3] George Washington Univ, Dept Elect & Comp Engn, Washington, DC 20052 USA
关键词
Deep learning; gradient data decomposition; streaming; principal component analysis;
D O I
10.1145/3577214
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
The movement of large quantities of data during the training of a deep neural network presents immense challenges for machine learning workloads, especially those based on future functional memories deployed to store network models. As the size of network models begins to vastly outstrip traditional silicon computing resources, functional memories based on flash, resistive switches, magnetic tunnel junctions, and other technologies can store these new ultra-large models. However, new approaches are then needed to minimize hardware overhead, especially on the movement and calculation of gradient information that cannot be efficiently contained in these new memory resources. To do this, we introduce streaming batch principal component analysis (SBPCA) as an update algorithm. Streaming batch principal component analysis uses stochastic power iterations to generate a stochastic rank-k approximation of the network gradient. We demonstrate that the low-rank updates produced by streaming batch principal component analysis can effectively train convolutional neural networks on a variety of common datasets, with performance comparable to standard mini-batch gradient descent. Our approximation is made in an expanded vector form that can efficiently be applied to the rows and columns of crossbars for array-level updates. These results promise improvements in the design of application-specific integrated circuits based around large vector-matrix multiplier memories.
引用
收藏
页数:24
相关论文
共 50 条
  • [1] LOW-RANK GRADIENT APPROXIMATION FOR MEMORY-EFFICIENT ON-DEVICE TRAINING OF DEEP NEURAL NETWORK
    Gooneratne, Mary
    Sim, Khe Chai
    Zadrazil, Petr
    Kabel, Andreas
    Beaufays, Francoise
    Motta, Giovanni
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 3017 - 3021
  • [2] Fast, memory-efficient low-rank approximation of SimRank
    Oseledets I.V.
    Ovchinnikov G.V.
    Katrutsa A.M.
    Journal of Complex Networks, 2017, 5 (01) : 111 - 126
  • [3] Low-Rank Gradient Descent
    Cosson, Romain
    Jadbabaie, Ali
    Makur, Anuran
    Reisizadeh, Amirhossein
    Shah, Devavrat
    IEEE Open Journal of Control Systems, 2023, 2 : 380 - 395
  • [4] Facto-CNN: Memory-Efficient CNN Training with Low-rank Tensor Factorization and Lossy Tensor Compression
    Lee, Seungtae
    Ko, Jonghwan
    Hong, Seokin
    ASIAN CONFERENCE ON MACHINE LEARNING, VOL 222, 2023, 222
  • [5] TT-TSDF: Memory-Efficient TSDF with Low-Rank Tensor Train Decomposition
    Boyko, Alexey, I
    Matrosov, Mikhail P.
    Oseledets, Ivan, V
    Tsetserukou, Dzmitry
    Ferrer, Gonzalo
    2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 10116 - 10121
  • [6] Adaptive Low-Rank Gradient Descent
    Jadbabaie, Ali
    Makur, Anuran
    Reisizadeh, Amirhossein
    2023 62ND IEEE CONFERENCE ON DECISION AND CONTROL, CDC, 2023, : 3315 - 3320
  • [7] Gradient Descent with Low-Rank Objective Functions
    Cosson, Romain
    Jadbabaie, Ali
    Makur, Anuran
    Reisizadeh, Amirhossein
    Shah, Devavrat
    2023 62ND IEEE CONFERENCE ON DECISION AND CONTROL, CDC, 2023, : 3309 - 3314
  • [8] Efficient Low-Rank Stochastic Gradient Descent Methods for Solving Semidefinite Programs
    Chen, Jianhui
    Yang, Tianbao
    Zhu, Shenghuo
    ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 33, 2014, 33 : 122 - 130
  • [9] Convergence of Gradient Descent for Low-Rank Matrix Approximation
    Pitaval, Renaud-Alexandre
    Dai, Wei
    Tirkkonen, Olav
    IEEE TRANSACTIONS ON INFORMATION THEORY, 2015, 61 (08) : 4451 - 4457
  • [10] A Memory-Efficient Hybrid Parallel Framework for Deep Neural Network Training
    Li, Dongsheng
    Li, Shengwei
    Lai, Zhiquan
    Fu, Yongquan
    Ye, Xiangyu
    Cai, Lei
    Qiao, Linbo
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2024, 35 (04) : 577 - 591