Low-Rank Gradient Descent for Memory-Efficient Training of Deep In-Memory Arrays

被引:0
|
作者
Huang, Siyuan [1 ]
Hoskins, Brian D. [2 ]
Daniels, Matthew W. [2 ]
Stiles, Mark D. [2 ]
Adam, Gina C. [3 ]
机构
[1] George Washington Univ, Dept Comp Sci, Washington, DC 20038 USA
[2] Natl Inst Stand & Technol, Gaithersburg, MD USA
[3] George Washington Univ, Dept Elect & Comp Engn, Washington, DC 20052 USA
关键词
Deep learning; gradient data decomposition; streaming; principal component analysis;
D O I
10.1145/3577214
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
The movement of large quantities of data during the training of a deep neural network presents immense challenges for machine learning workloads, especially those based on future functional memories deployed to store network models. As the size of network models begins to vastly outstrip traditional silicon computing resources, functional memories based on flash, resistive switches, magnetic tunnel junctions, and other technologies can store these new ultra-large models. However, new approaches are then needed to minimize hardware overhead, especially on the movement and calculation of gradient information that cannot be efficiently contained in these new memory resources. To do this, we introduce streaming batch principal component analysis (SBPCA) as an update algorithm. Streaming batch principal component analysis uses stochastic power iterations to generate a stochastic rank-k approximation of the network gradient. We demonstrate that the low-rank updates produced by streaming batch principal component analysis can effectively train convolutional neural networks on a variety of common datasets, with performance comparable to standard mini-batch gradient descent. Our approximation is made in an expanded vector form that can efficiently be applied to the rows and columns of crossbars for array-level updates. These results promise improvements in the design of application-specific integrated circuits based around large vector-matrix multiplier memories.
引用
收藏
页数:24
相关论文
共 50 条
  • [41] Hybrid In-memory Computing Architecture for the Training of Deep Neural Networks
    Joshi, Vinay
    He, Wangxin
    Seo, Jae-sun
    Rajendran, Bipin
    2021 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2021,
  • [42] Non-convex projected gradient descent for generalized low-rank tensor regression
    Chen, Han
    Raskutti, Garvesh
    Yuan, Ming
    Journal of Machine Learning Research, 2019, 20
  • [43] Non-Convex Projected Gradient Descent for Generalized Low-Rank Tensor Regression
    Chen, Han
    Raskutti, Garvesh
    Yuan, Ming
    JOURNAL OF MACHINE LEARNING RESEARCH, 2019, 20
  • [44] BYZANTINE-ROBUST STOCHASTIC GRADIENT DESCENT FOR DISTRIBUTED LOW-RANK MATRIX COMPLETION
    He, Xuechao
    Ling, Qing
    Chen, Tianyi
    2019 IEEE DATA SCIENCE WORKSHOP (DSW), 2019, : 322 - 326
  • [45] Sequence Discriminative Training for Low-Rank Deep Neural Networks
    Tachioka, Yuuki
    Watanabe, Shinji
    Le Roux, Jonathan
    Hershey, John R.
    2014 IEEE GLOBAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (GLOBALSIP), 2014, : 572 - 576
  • [46] Design of low-power memory-efficient viterbi decoder
    Chen, Lupin
    He, Jinjin
    Wang, Zhongfeng
    2007 IEEE WORKSHOP ON SIGNAL PROCESSING SYSTEMS, VOLS 1 AND 2, 2007, : 132 - 135
  • [47] A Scalable Near-Memory Architecture for Training Deep Neural Networks on Large In-Memory Datasets
    Schuiki, Fabian
    Schaffner, Michael
    Gurkaynak, Frank K.
    Benini, Luca
    IEEE TRANSACTIONS ON COMPUTERS, 2019, 68 (04) : 484 - 497
  • [48] Efficient SVM training using low-rank kernel representations
    Fine, S
    Scheinberg, K
    JOURNAL OF MACHINE LEARNING RESEARCH, 2002, 2 (02) : 243 - 264
  • [49] Pruned RNN-T for fast, memory-efficient ASR training
    Kuang, Fangjun
    Guo, Liyong
    Kang, Wei
    Lin, Long
    Luo, Mingshuang
    Yao, Zengwei
    Povey, Daniel
    INTERSPEECH 2022, 2022, : 2068 - 2072
  • [50] Memory-Efficient Deep Learning for Botnet Attack Detection in IoT Networks
    Popoola, Segun I.
    Adebisi, Bamidele
    Ande, Ruth
    Hammoudeh, Mohammad
    Atayero, Aderemi A.
    ELECTRONICS, 2021, 10 (09)