Hardware-aware approach to deep neural network optimization

被引:3
|
作者
Li, Hengyi [1 ]
Meng, Lin [2 ]
机构
[1] Ritsumeikan Univ, Res Org Sci & Technol, 1-1-1 Noji Higashi, Kusatsu, Shiga 5258577, Japan
[2] Ritsumeikan Univ, Coll Sci & Engn, 1-1-1 Noji Higashi, Kusatsu, Shiga 5258577, Japan
关键词
DNNs; Optimization; Hardware-aware approach; LWPolar; Polar_HSPG; IHSOpti; DESIGN; FRAMEWORK;
D O I
10.1016/j.neucom.2023.126808
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNs) have been a pivotal technology in a myriad of fields, boasting remarkable achievements. Nevertheless, their substantial workload and inherent redundancies pose ongoing challenges for both practitioners and academia. While numerous researchers endeavor to optimize DNNs, the inherent parallelism features of hardware are generally underutilized, resulting in inefficient use of hardware resources. To address this deficit, the paper unveils a hardware-aware mechanism, IHSOpti, which incorporates hardware characteristics with software algorithms for DNN optimization. IHSOpti endeavors to exploit the full potential of modern hardware parallelism, with significant emphasis on pipelining mechanisms. Specifically, IHSOpti for-mulates an advanced sparse training algorithm Polar_HSPG which incorporates the newly-proposed layer-wise refined polarization regularizer (LWPolar), grounded on the half-space project gradient (HSPG). Subsequently, IHSOpti pioneeringly introduces the residual strategy for optimizing the layer-level redundancies of neural networks, capitalizing on the pipelining attributes inherent in current hardware. Experimental findings demonstrate that IHSOpti attains outstanding pruning ratios in both parameters and FLOPs. Specifically, IHSOpti achieves up to 96.90% and 82.73% pruning ratios with the accuracy of 93.34% for VGGBN, 97.69% and 95.24% pruning ratios with the accuracy of 94.69% for ResNet, 98.07% and 97.80% pruning ratios with the accuracy of 95.73% for the cutting-edge network RegNet, respectively. Notably, the running efficiency exhibits remarkable improvements with accelerations ranging from 3.63x to 8.20x for CPUs and 1.22x to 2.25x for GPUs, respectively. These outcomes surpass the latest advances in the field. Through the incorporation of specific hardware characteristics, IHSOpti provides a comprehensive and effective approach to harness the intrinsic parallelism of contemporary hardware platforms for DNNs.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] SqueezeNext: Hardware-Aware Neural Network Design
    Gholami, Amir
    Kwon, Kiseok
    Wu, Bichen
    Tai, Zizheng
    Yue, Xiangyu
    Jin, Peter
    Zhao, Sicheng
    Keutzer, Kurt
    [J]. PROCEEDINGS 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2018, : 1719 - 1728
  • [2] Exploring Quantization and Mapping Synergy in Hardware-Aware Deep Neural Network Accelerators
    Klhufek, Jan
    Safar, Miroslav
    Mrazek, Vojtech
    Vasicek, Zdenek
    Sekanina, Lukas
    [J]. 2024 27TH INTERNATIONAL SYMPOSIUM ON DESIGN & DIAGNOSTICS OF ELECTRONIC CIRCUITS & SYSTEMS, DDECS, 2024, : 1 - 6
  • [3] Hardware-Aware Softmax Approximation for Deep Neural Networks
    Geng, Xue
    Lin, Jie
    Zhao, Bin
    Kong, Anmin
    Aly, Mohamed M. Sabry
    Chandrasekhar, Vijay
    [J]. COMPUTER VISION - ACCV 2018, PT IV, 2019, 11364 : 107 - 122
  • [4] Hardware-Aware Quantization for Multiplierless Neural Network Controllers
    Habermann, Tobias
    Kuehle, Jonas
    Kumm, Martin
    Volkova, Anastasia
    [J]. 2022 IEEE ASIA PACIFIC CONFERENCE ON CIRCUITS AND SYSTEMS, APCCAS, 2022, : 541 - 545
  • [5] HAO: Hardware-aware Neural Architecture Optimization for Efficient Inference
    Dong, Zhen
    Gao, Yizhao
    Huang, Qijing
    Wawrzynek, John
    So, Hayden K. H.
    Keutzer, Kurt
    [J]. 2021 IEEE 29TH ANNUAL INTERNATIONAL SYMPOSIUM ON FIELD-PROGRAMMABLE CUSTOM COMPUTING MACHINES (FCCM 2021), 2021, : 50 - 59
  • [6] Hardware-Aware Model of Sigma-Delta Cellular Neural Network
    Aomori, Hisashi
    Naito, Yuki
    Otake, Tsuyoshi
    Takahashi, Nobuaki
    Matsuda, Ichiro
    Itoh, Susumu
    Tanaka, Mamoru
    [J]. 2009 EUROPEAN CONFERENCE ON CIRCUIT THEORY AND DESIGN, VOLS 1 AND 2, 2009, : 311 - +
  • [7] Quantized rewiring: hardware-aware training of sparse deep neural networks
    Petschenig, Horst
    Legenstein, Robert
    [J]. NEUROMORPHIC COMPUTING AND ENGINEERING, 2023, 3 (02):
  • [8] On Hardware-Aware Design and Optimization of Edge Intelligence
    Huai, Shuo
    Kong, Hao
    Luo, Xiangzhong
    Liu, Di
    Subramaniam, Ravi
    Makaya, Christian
    Lin, Qian
    Liu, Weichen
    [J]. IEEE DESIGN & TEST, 2023, 40 (06) : 149 - 162
  • [9] Hardware-Aware Analysis and Optimization of Stable Fluids
    Kim, Theodore
    [J]. I3D 2008: SYMPOSIUM ON INTERACTIVE 3D GRAPHICS AND GAMES, PROCEEDINGS, 2008, : 99 - 106
  • [10] HFP: Hardware-Aware Filter Pruning for Deep Convolutional Neural Networks Acceleration
    Yu, Fang
    Han, Chuanqi
    Wang, Pengcheng
    Huang, Ruoran
    Huang, Xi
    Cui, Li
    [J]. 2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 255 - 262