Partitioning Sparse Deep Neural Networks for Scalable Training and Inference

被引:6
|
作者
Demirci, Gunduz Vehbi [1 ]
Ferhatosmanoglu, Hakan [1 ]
机构
[1] Univ Warwick, Coventry, W Midlands, England
关键词
Scalable Deep Learning; Sparse Deep Neural Networks; Distributed Stochastic Gradient Descent; Hypergraph Partitioning; Sparse Matrix Vector Multiplication;
D O I
10.1145/3447818.3460372
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
The state-of-the-art deep neural networks (DNNs) have significant computational and data management requirements. The size of both training data and models continue to increase. Sparsification and pruning methods are shown to be effective in removing a large fraction of connections in DNNs. The resulting sparse networks present unique challenges to further improve the computational efficiency of training and inference in deep learning. Both the feed-forward (inference) and backpropagation steps in stochastic gradient descent (SGD) algorithm for training sparse DNNs involve consecutive sparse matrix-vector multiplications (SpMVs). We first introduce a distributed-memory parallel SpMV-based solution for the SGD algorithm to improve its scalability. The parallelization approach is based on row-wise partitioning of weight matrices that represent neuron connections between consecutive layers. We then propose a novel hypergraph model for partitioning weight matrices to reduce the total communication volume and ensure computational load-balance among processors. Experiments performed on sparse DNNs demonstrate that the proposed solution is highly efficient and scalable. By utilizing the proposed matrix partitioning scheme, the performance of our solution is further improved significantly.
引用
收藏
页码:254 / 265
页数:12
相关论文
共 50 条
  • [31] CEST: Computation-Efficient N:M Sparse Training for Deep Neural Networks
    Fang, Chao
    Sun, Wei
    Zhou, Aojun
    Wang, Zhongfeng
    [J]. 2023 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION, DATE, 2023,
  • [32] TRIM: A Design Space Exploration Model for Deep Neural Networks Inference and Training Accelerators
    Qi, Yangjie
    Zhang, Shuo
    Taha, Tarek M.
    [J]. IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2023, 42 (05) : 1648 - 1661
  • [33] A Tandem Learning Rule for Effective Training and Rapid Inference of Deep Spiking Neural Networks
    Wu, Jibin
    Chua, Yansong
    Zhang, Malu
    Li, Guoqi
    Li, Haizhou
    Tan, Kay Chen
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (01) : 446 - 460
  • [34] Sparse solution in training artificial neural networks
    Giustolisi, O
    [J]. NEUROCOMPUTING, 2004, 56 : 285 - 304
  • [35] Training Neural Networks with Fixed Sparse Masks
    Sung, Yi-Lin
    Nair, Varun
    Raffel, Colin
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [36] Scalable Training of Inference Networks for Gaussian-Process Models
    Shi, Jiaxin
    Khan, Mohammad Emtiyaz
    Zhu, Jun
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [37] Calibration-Aided Edge Inference Offloading via Adaptive Model Partitioning of Deep Neural Networks
    Pacheco, Roberto G.
    Couto, Rodrigo S.
    Simeone, Osvaldo
    [J]. IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC 2021), 2021,
  • [38] Sparse Deep Neural Networks for Embedded Intelligence
    Bi, Jia
    Gunn, Steve R.
    [J]. 2018 IEEE 30TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI), 2018, : 30 - 38
  • [39] Scaling for edge inference of deep neural networks
    Xiaowei Xu
    Yukun Ding
    Sharon Xiaobo Hu
    Michael Niemier
    Jason Cong
    Yu Hu
    Yiyu Shi
    [J]. Nature Electronics, 2018, 1 : 216 - 222
  • [40] Learning Sparse Patterns in Deep Neural Networks
    Wen, Weijing
    Yang, Fan
    Su, Yangfeng
    Zhou, Dian
    Zeng, Xuan
    [J]. 2019 IEEE 13TH INTERNATIONAL CONFERENCE ON ASIC (ASICON), 2019,