Towards Energy Efficient DNN accelerator via Sparsified Gradual Knowledge Distillation

被引:1
|
作者
Karimzadeh, Foroozan [1 ]
Raychowdhury, Arijit [1 ]
机构
[1] Georgia Inst Technol, Sch Elect & Comp Engn, Atlanta, GA 30332 USA
关键词
DNN model; Knowledge distillation; DNN compression; quantization; low bit precision; NETWORK;
D O I
10.1109/VLSI-SoC54400.2022.9939619
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Artificial intelligence (AI) is becoming increasingly popular in many applications. However, the computation cost of deep neural network (DNN) , which is a powerful form of AI, calls for efficient DNN compression technique to make energy efficient networks. In this paper, we proposed SKG, a method to jointly sparsify and quantize DNN models to ultra-low bit-precision using Knowledge Distillation and gradual quantization (SKG). We demonstrated that our method can preserve the accuracy more than 20% for uniform quantization with 2 bit-width compared to the baseline methods on ImageNet and ResNet-18. In addition, our method can achieve up to 2.7x lower energy consumption using compute-in-memory (CIM) architecture compared to a traditional 65nm CMOS architecture for both pruned and unpruned network during inference and eventually enabling using DNN models on resource constrained edge devices.
引用
收藏
页数:6
相关论文
共 50 条
  • [1] Towards CIM-friendly and Energy-Efficient DNN Accelerator via Bit-level Sparsity
    Karimzadeh, Foroozan
    Raychowdhury, Arijit
    [J]. PROCEEDINGS OF THE 2022 IFIP/IEEE 30TH INTERNATIONAL CONFERENCE ON VERY LARGE SCALE INTEGRATION (VLSI-SOC), 2022,
  • [2] Cyclic distillation - towards energy efficient binary distillation
    Kiss, Anton A.
    Landaeta, Servando J. Flores
    Zondervan, Edwin
    [J]. 22 EUROPEAN SYMPOSIUM ON COMPUTER AIDED PROCESS ENGINEERING, 2012, 30 : 697 - 701
  • [3] Towards an Efficient Accelerator for DNN-based Remote Sensing Image Segmentation on FPGAs
    Liu, Shuanglong
    Luk, Wayne
    [J]. 2019 29TH INTERNATIONAL CONFERENCE ON FIELD-PROGRAMMABLE LOGIC AND APPLICATIONS (FPL), 2019, : 187 - 193
  • [4] Efficient Crowd Counting via Dual Knowledge Distillation
    Wang, Rui
    Hao, Yixue
    Hu, Long
    Li, Xianzhi
    Chen, Min
    Miao, Yiming
    Humar, Iztok
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 569 - 583
  • [5] Efficient Biomedical Instance Segmentation via Knowledge Distillation
    Liu, Xiaoyu
    Hu, Bo
    Huang, Wei
    Zhang, Yueyi
    Xiong, Zhiwei
    [J]. MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2022, PT IV, 2022, 13434 : 14 - 24
  • [6] PIMCA: A Programmable In-Memory Computing Accelerator for Energy-Efficient DNN Inference
    Zhang, Bo
    Yin, Shihui
    Kim, Minkyu
    Saikia, Jyotishman
    Kwon, Soonwan
    Myung, Sungmeen
    Kim, Hyunsoo
    Kim, Sang Joon
    Seo, Jae-Sun
    Seok, Mingoo
    [J]. IEEE JOURNAL OF SOLID-STATE CIRCUITS, 2023, 58 (05) : 1436 - 1449
  • [7] Energy Efficient Boosting of GEMM Accelerators for DNN via Reuse
    Cicek, Nihat Mert
    Shen, Xipeng
    Ozturk, Ozcan
    [J]. ACM TRANSACTIONS ON DESIGN AUTOMATION OF ELECTRONIC SYSTEMS, 2022, 27 (05)
  • [8] QeiHaN: An Energy-Efficient DNN Accelerator that Leverages Log Quantization in NDP Architectures
    Khabbazan, Bahareh
    Riera, Marc
    Gonzalez, Antonio
    [J]. 2023 32ND INTERNATIONAL CONFERENCE ON PARALLEL ARCHITECTURES AND COMPILATION TECHNIQUES, PACT, 2023, : 325 - 326
  • [9] Towards Continual Knowledge Graph Embedding via Incremental Distillation
    Liu, Jiajun
    Ke, Wenjun
    Wang, Peng
    Shang, Ziyu
    Gao, Jinhua
    Li, Guozheng
    Ji, Ke
    Liu, Yanhe
    [J]. THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 8, 2024, : 8759 - 8768
  • [10] Towards General and Fast Video Derain via Knowledge Distillation
    Cai, Defang
    Mu, Pan
    Chan, Sixian
    Shao, Zhanpeng
    Bai, Cong
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 1949 - 1954