Eager Pruning: Algorithm and Architecture Support for Fast Training of Deep Neural Networks

被引:41
|
作者
Zhang, Jiaqi [1 ]
Chen, Xiangru [1 ]
Song, Mingcong [1 ]
Li, Tao [1 ]
机构
[1] Univ Florida, Dept Elect & Comp Engn, Gainesville, FL 32611 USA
基金
美国国家科学基金会;
关键词
Neural Network Training; Neural Network Pruning; Software-Hardware Co-Design;
D O I
10.1145/3307650.3322263
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Today's big and fast data and the changing circumstance require fast training of Deep Neural Networks (DNN) in various applications. However, training a DNN with tons of parameters involves intensive computation. Enlightened by the fact that redundancy exists in DNNs and the observation that the ranking of the significance of the weights changes slightly during training, we propose Eager Pruning, which speeds up DNN training by moving pruning to an early stage. Eager Pruning is supported by an algorithm and architecture co-design. The proposed algorithm dictates the architecture to identify and prune insignificant weights during training without accuracy loss. A novel architecture is designed to transform the reduced training computation into performance improvement. Our proposed Eager Pruning system gains an average of 1.91x speedup over state-of-the-art hardware accelerator and 6.31x energy-efficiency over Nvidia GPUs.
引用
收藏
页码:292 / 303
页数:12
相关论文
共 50 条
  • [1] DyGNN: Algorithm and Architecture Support of Dynamic Pruning for Graph Neural Networks
    Chen, Cen
    Li, Kenli
    Zou, Xiaofeng
    Li, Yangfan
    [J]. 2021 58TH ACM/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2021, : 1201 - 1206
  • [2] A fast adaptive algorithm for training deep neural networks
    Gui, Yangting
    Li, Dequan
    Fang, Runyue
    [J]. APPLIED INTELLIGENCE, 2023, 53 (04) : 4099 - 4108
  • [3] A fast adaptive algorithm for training deep neural networks
    Yangting Gui
    Dequan Li
    Runyue Fang
    [J]. Applied Intelligence, 2023, 53 : 4099 - 4108
  • [4] Fast Convex Pruning of Deep Neural Networks
    Aghasi, Alireza
    Abdi, Afshin
    Romberg, Justin
    [J]. SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2020, 2 (01): : 158 - 188
  • [5] Fast-M Adversarial Training Algorithm for Deep Neural Networks
    Ma, Yu
    An, Dou
    Gu, Zhixiang
    Lin, Jie
    Liu, Weiyu
    [J]. APPLIED SCIENCES-BASEL, 2024, 14 (11):
  • [6] A fast training algorithm for neural networks
    Bilski, J
    Rutkowski, L
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-ANALOG AND DIGITAL SIGNAL PROCESSING, 1998, 45 (06): : 749 - 753
  • [7] The Impact of Architecture on the Deep Neural Networks Training
    Rozycki, Pawel
    Kolbusz, Janusz
    Malinowski, Aleksander
    Wilamowski, Bogdan
    [J]. 2019 12TH INTERNATIONAL CONFERENCE ON HUMAN SYSTEM INTERACTION (HSI), 2019, : 41 - 46
  • [8] An algorithm for fast convergence in training neural networks
    Wilamowski, BM
    Iplikci, S
    Kaynak, O
    Efe, MÖ
    [J]. IJCNN'01: INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-4, PROCEEDINGS, 2001, : 1778 - 1782
  • [9] A Fast Compressed Hardware Architecture for Deep Neural Networks
    Ansari, Anaam
    Shelton, Allen
    Ogunfunmi, Tokunbo
    Panchbhaiyye, Vineet
    [J]. 2022 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS 22), 2022, : 370 - 374
  • [10] FAST TRAINING OF DEEP NEURAL NETWORKS FOR SPEECH RECOGNITION
    Gong, Guojing
    Kingsbury, Brian
    Yang, Chih-Chieh
    Liu, Tianyi
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 6884 - 6888