Energy-efficient Amortized Inference with Cascaded Deep Classifiers

被引:0
|
作者
Guan, Jiaqi [1 ,2 ]
Liu, Yang [2 ]
Liu, Qiang [3 ]
Peng, Jian [2 ]
机构
[1] Tsinghua Univ, Beijing, Peoples R China
[2] Univ Illinois, Urbana, IL USA
[3] Univ Texas Austin, Austin, TX 78712 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks have been remarkable successful in various AI tasks but often cast high computation and energy cost for energy-constrained applications such as mobile sensing. We address this problem by proposing a novel framework that optimizes the prediction accuracy and energy cost simultaneously, thus enabling effective cost-accuracy trade-off at test time. In our framework, each data instance is pushed into a cascade of deep neural networks with increasing sizes, and a selection module is used to sequentially determine when a sufficiently accurate classifier can be used for this data instance. The cascade of neural networks and the selection module are jointly trained in an end-to-end fashion by the REINFORCE algorithm to optimize a trade-off between the computational cost and the predictive accuracy. Our method is able to simultaneously improve the accuracy and efficiency by learning to assign easy instances to fast yet sufficiently accurate classifiers to save computation and energy cost, while assigning harder instances to deeper and more powerful classifiers to ensure satisfiable accuracy. Moreover, we demonstrate our method's effectiveness with extensive experiments on CIFAR-10/100, ImageNet32x32 and original ImageNet dataset.
引用
收藏
页码:2184 / 2190
页数:7
相关论文
共 50 条
  • [1] Energy-efficient deep learning inference on edge devices
    Daghero, Francesco
    Pagliari, Daniele Jahier
    Poncino, Massimo
    [J]. HARDWARE ACCELERATOR SYSTEMS FOR ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING, 2021, 122 : 247 - 301
  • [2] Energy-Efficient Neuromorphic Classifiers
    Marti, Daniel
    Rigotti, Mattia
    Seok, Mingoo
    Fusi, Stefano
    [J]. NEURAL COMPUTATION, 2016, 28 (10) : 2011 - 2044
  • [3] PIE: A Pipeline Energy-efficient Accelerator for Inference Process in Deep Neural Networks
    Zhao, Yangyang
    Yu, Qi
    Zhou, Xuda
    Zhou, Xuehai
    Wang, Chao
    Li, Xi
    [J]. 2016 IEEE 22ND INTERNATIONAL CONFERENCE ON PARALLEL AND DISTRIBUTED SYSTEMS (ICPADS), 2016, : 1067 - 1074
  • [4] Energy-Efficient Embedded Inference of SVMs on FPGA
    Elgawi, Osman
    Mutawa, A. M.
    Ahmad, Afaq
    [J]. 2019 IEEE COMPUTER SOCIETY ANNUAL SYMPOSIUM ON VLSI (ISVLSI 2019), 2019, : 165 - 169
  • [5] Energy-Efficient Approximate Edge Inference Systems
    Ghosh, Soumendu Kumar
    Raha, Arnab
    Raghunathan, Vijay
    [J]. ACM TRANSACTIONS ON EMBEDDED COMPUTING SYSTEMS, 2023, 22 (04)
  • [6] Energy-efficient cooperative inference via adaptive deep neural network splitting at the edge
    Labriji, Ibtissam
    Merluzzi, Mattia
    Airod, Fatima Ezzahra
    Strinati, Emilio Calvanese
    [J]. ICC 2023-IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS, 2023, : 1712 - 1717
  • [7] TIE: Energy-efficient Tensor Train-based Inference Engine for Deep Neural Network
    Deng, Chunhua
    Sun, Fangxuan
    Qian, Xuehai
    Lin, Jun
    Wang, Zhongfeng
    Yuan, Bo
    [J]. PROCEEDINGS OF THE 2019 46TH INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE (ISCA '19), 2019, : 264 - 277
  • [8] Poster Abstract: MicroBrain: Compressing Deep Neural Networks for Energy-efficient Visual Inference Service
    Ge, Shiming
    Luo, Zhao
    Ye, Qiting
    Zhang, Xiao-Yu
    [J]. 2017 IEEE CONFERENCE ON COMPUTER COMMUNICATIONS WORKSHOPS (INFOCOM WKSHPS), 2017, : 1000 - 1001
  • [9] SONIC: A Sparse Neural Network Inference Accelerator with Silicon Photonics for Energy-Efficient Deep Learning
    Sunny, Febin
    Nikdast, Mandi
    Pasricha, Sudeep
    [J]. 27TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE, ASP-DAC 2022, 2022, : 214 - 219
  • [10] Energy-Efficient Bayesian Inference Using Bitstream Computing
    Khoram, Soroosh
    Daruwalla, Kyle
    Lipasti, Mikko
    [J]. IEEE COMPUTER ARCHITECTURE LETTERS, 2023, 22 (01) : 37 - 40