Compressing DMA Engine: Leveraging Activation Sparsity for Training Deep Neural Networks

被引:107
|
作者
Rhu, Minsoo [1 ]
O'Connor, Mike [2 ]
Chatterjee, Niladrish [2 ]
Pool, Jeff [2 ]
Kwon, Youngeun [1 ]
Keckler, Stephen W. [2 ]
机构
[1] POSTECH, Pohang, South Korea
[2] NVIDIA, Santa Clara, CA USA
关键词
D O I
10.1109/HPCA.2018.00017
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Popular deep learning frameworks require users to fine-tune their memory usage so that the training data of a deep neural network (DNN) fits within the GPU physical memory. Prior work tries to address this restriction by virtualizing the memory usage of DNNs, enabling both CPU and GPU memory to be utilized for memory allocations. Despite its merits, virtualizing memory can incur significant performance overheads when the time needed to copy data back and forth from CPU memory is higher than the latency to perform DNN computations. We introduce a high-performance virtualization strategy based on a "compressing DMA engine" (cDMA) that drastically reduces the size of the data structures that are targeted for CPU-side allocations. The cDMA engine offers an average 2.6x (maximum 13.8x) compression ratio by exploiting the sparsity inherent in offloaded data, improving the performance of virtualized DNNs by an average 53% (maximum 79%) when evaluated on an NVIDIA Titan Xp.
引用
收藏
页码:78 / 91
页数:14
相关论文
共 50 条
  • [1] Compressing Low Precision Deep Neural Networks Using Sparsity-Induced Regularization in Ternary Networks
    Faraone, Julian
    Fraser, Nicholas
    Gambardella, Giulio
    Blott, Michaela
    Leong, Philip H. W.
    NEURAL INFORMATION PROCESSING (ICONIP 2017), PT II, 2017, 10635 : 393 - 404
  • [2] Addressing Sparsity in Deep Neural Networks
    Zhou, Xuda
    Du, Zidong
    Zhang, Shijin
    Zhang, Lei
    Lan, Huiying
    Liu, Shaoli
    Li, Ling
    Guo, Qi
    Chen, Tianshi
    Chen, Yunji
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2019, 38 (10) : 1858 - 1871
  • [3] Compressing Deep Neural Networks for Recognizing Places
    Saha, Soham
    Varma, Girish
    Jawahar, C. V.
    PROCEEDINGS 2017 4TH IAPR ASIAN CONFERENCE ON PATTERN RECOGNITION (ACPR), 2017, : 352 - 357
  • [4] Sparsity in deep learning: Pruning and growth for efficient inference and training in neural networks
    Hoefler, Torsten
    Alistarh, Dan
    Ben-Nun, Tal
    Dryden, Nikoli
    Peste, Alexandra
    Journal of Machine Learning Research, 2021, 22
  • [5] Sparsity in Deep Learning: Pruning and growth for efficient inference and training in neural networks
    Hoefler, Torsten
    Alistarh, Dan
    Ben-Nun, Tal
    Dryden, Nikoli
    Peste, Alexandra
    JOURNAL OF MACHINE LEARNING RESEARCH, 2021, 23
  • [6] Sparse ReRAM Engine: Joint Exploration of Activation and Weight Sparsity in Compressed Neural Networks
    Yang, Tzu-Hsien
    Cheng, Hsiang-Yun
    Yang, Chia-Lin
    Tseng, I-Ching
    Hu, Han-Wen
    Chang, Hung-Sheng
    Li, Hsiang-Pang
    PROCEEDINGS OF THE 2019 46TH INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE (ISCA '19), 2019, : 236 - 249
  • [7] Parallax: Sparsity-aware Data Parallel Training of Deep Neural Networks
    Kim, Soojeong
    Yu, Gyeong-In
    Park, Hojin
    Cho, Sungwoo
    Jeong, Eunji
    Ha, Hyeonmin
    Lee, Sanha
    Jeong, Joo Seong
    Chun, Byung-Gon
    PROCEEDINGS OF THE FOURTEENTH EUROSYS CONFERENCE 2019 (EUROSYS '19), 2019,
  • [8] Accelerating Deep Neural Networks via Semi-Structured Activation Sparsity
    Grimaldi, Matteo
    Ganji, Darshan C.
    Lazarevich, Ivan
    Sah, Sudhakar
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, : 1171 - 1180
  • [9] On the Impact of the Activation Function on Deep Neural Networks Training
    Hayou, Soufiane
    Doucet, Arnaud
    Rousseau, Judith
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [10] Learning Structured Sparsity in Deep Neural Networks
    Wen, Wei
    Wu, Chunpeng
    Wang, Yandan
    Chen, Yiran
    Li, Hai
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29