Impact of Mixed Precision Techniques on Training and Inference Efficiency of Deep Neural Networks

被引:2
|
作者
Doerrich, Marion [1 ]
Fan, Mingcheng [1 ]
Kist, Andreas M. [1 ]
机构
[1] Friedrich Alexander Univ Erlangen Nurnberg, Dept Artificial Intelligence Biomed Engn, D-91052 Erlangen, Germany
关键词
Deep learning; green AI; energy efficiency; mixed precision training; quantization; edge TPU;
D O I
10.1109/ACCESS.2023.3284388
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In the deep learning community, increasingly large models are being developed, leading to rapidly growing computational costs and energy costs. Recently, a new trend has been arising, advocating that researchers should also report the energy efficiency besides their model's performance in their papers. Previous research has shown that reduced precision can be helpful to improve energy efficiency. Based on this finding, we propose a simple practice to effectively improve the energy efficiency of training and inference, i.e., training the model with mixed precision and deploying it on Edge TPUs. We evaluated its effectiveness by comparing the speed-up of a state-of-the-art semantic segmentation architecture with respect to different typical usage scenarios, including using different devices, deep learning frameworks, model sizes, and batch sizes. Our results show that enabled mixed precision can gain up to a $1.9\times $ speedup compared to the most common and default float32 data type on GPUs. Deploying the models on Edge TPU further boosted the inference by a factor of 6. Our approach allows researchers to accelerate their training and inference procedures without jeopardizing the model's accuracy, meanwhile reducing energy consumption and electricity cost easily without changing their model architecture or retraining. Furthermore, our approach is helpful in reducing the carbon footprint used to train and deploy the neural network and thus has a positive effect on environmental resources.
引用
收藏
页码:57627 / 57634
页数:8
相关论文
共 50 条
  • [1] POSITNN: TRAINING DEEP NEURAL NETWORKS WITH MIXED LOW-PRECISION POSIT
    Raposo, Goncalo
    Tomas, Pedro
    Roma, Nuno
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 7908 - 7912
  • [2] Mixed Precision Weight Networks: Training Neural Networks with Varied Precision Weights
    Fuengfusin, Ninnart
    Tamukoh, Hakaru
    [J]. NEURAL INFORMATION PROCESSING (ICONIP 2018), PT II, 2018, 11302 : 614 - 623
  • [3] Mixed-precision architecture based on computational memory for training deep neural networks
    Nandakumar, S. R.
    Le Gallo, Manuel
    Boybat, Irem
    Rajendran, Bipin
    Sebastian, Abu
    Eleftheriou, Evangelos
    [J]. 2018 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2018,
  • [4] Evaluating the Impact of Mixed-Precision on Fault Propagation for Deep Neural Networks on GPUs
    Dos Santos, Fernando Fernandes
    Rech, Paolo
    Kritikakou, Angeliki
    Sentieys, Olivier
    [J]. 2022 IEEE COMPUTER SOCIETY ANNUAL SYMPOSIUM ON VLSI (ISVLSI 2022), 2022, : 327 - 327
  • [5] Applying Lightweight Soft Error Mitigation Techniques to Embedded Mixed Precision Deep Neural Networks
    Abich, Geancarlo
    Gava, Jonas
    Garibotti, Rafael
    Reis, Ricardo
    Ost, Luciano
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2021, 68 (11) : 4772 - 4782
  • [6] Partitioning Sparse Deep Neural Networks for Scalable Training and Inference
    Demirci, Gunduz Vehbi
    Ferhatosmanoglu, Hakan
    [J]. PROCEEDINGS OF THE 2021 ACM INTERNATIONAL CONFERENCE ON SUPERCOMPUTING, ICS 2021, 2021, : 254 - 265
  • [7] The Impact of Architecture on the Deep Neural Networks Training
    Rozycki, Pawel
    Kolbusz, Janusz
    Malinowski, Aleksander
    Wilamowski, Bogdan
    [J]. 2019 12TH INTERNATIONAL CONFERENCE ON HUMAN SYSTEM INTERACTION (HSI), 2019, : 41 - 46
  • [8] Hardware for Quantized Mixed-Precision Deep Neural Networks
    Rios, Andres
    Nava, Patricia
    [J]. PROCEEDINGS OF THE 2022 15TH IEEE DALLAS CIRCUITS AND SYSTEMS CONFERENCE (DCAS 2022), 2022,
  • [9] Better schedules for low precision training of deep neural networks
    Cameron R. Wolfe
    Anastasios Kyrillidis
    [J]. Machine Learning, 2024, 113 : 3569 - 3587
  • [10] Better schedules for low precision training of deep neural networks
    Wolfe, Cameron R.
    Kyrillidis, Anastasios
    [J]. MACHINE LEARNING, 2024, 113 (06) : 3569 - 3587