Fine-grained Hardware Acceleration for Efficient Batteryless Intermittent Inference on the Edge

被引:4
|
作者
Caronti, Luca [1 ]
Akhunov, Khakim [1 ]
Nardello, Matteo [1 ]
Yildirim, Kasim Sinan [1 ]
Brunelli, Davide [1 ]
机构
[1] Univ Trento, Via Sommarive 9, I-38123 Trento, TN, Italy
关键词
Intermittent computing; convolutional neural networks; edge computing; energy harvesting; hardware accelerator; checkpointing;
D O I
10.1145/3608475
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Backing up the intermediate results of hardware-accelerated deep inference is crucial to ensure the progress of execution on batteryless computing platforms. However, hardware accelerators in low-power AI platforms only support the one-shot atomic execution of one neural network inference without any backups. This article introduces a new toolchain for MAX78000, which is a brand-new microcontroller with a hardware-based convolutional neural network (CNN) accelerator. Our toolchain converts any MAX78000-compatible neural network into an intermittently executable form. The toolchain enables finer checkpoint granularity on the MAX78000 CNN accelerator, allowing for backups of any intermediate neural network layer output. Based on the layer-by-layer CNN execution, we propose a new backup technique that performs only necessary (urgent) checkpoints. The method involves the batteryless system switching to ultra-low-power mode while charging, saving intermediate results only when input power is lower than ultra-low-power mode energy consumption. By avoiding unnecessary memory transfer, the proposed solution increases the inference throughput by 1.9x for simulation and by 1.2x for real-world setup compared to the coarse-grained baseline execution.
引用
收藏
页数:19
相关论文
共 50 条
  • [21] ENFIRE: A Spatio-Temporal Fine-Grained Reconfigurable Hardware
    Qian, Wenchao
    Babecki, Christopher
    Karam, Robert
    Paul, Somnath
    Bhunia, Swarup
    IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, 2017, 25 (01) : 177 - 188
  • [22] Realistic acceleration of neural networks with fine-grained tensor decomposition
    Lv, Rui
    Wang, Dingheng
    Zheng, Jiangbin
    Xie, Yefan
    Yang, Zhao-Xu
    NEUROCOMPUTING, 2022, 512 : 52 - 68
  • [23] Realizing Fine-Grained Inference of AS Path With a Generative Measurable Process
    Li, Xionglve
    Zhou, Tongqing
    Cai, Zhiping
    Su, Jinshu
    IEEE-ACM TRANSACTIONS ON NETWORKING, 2023, 31 (06) : 3112 - 3127
  • [24] An FPGA Overlay for CNN Inference with Fine-grained Flexible Parallelism
    Choudhury, Ziaul
    Shrivastava, Shashwat
    Ramapantulu, Lavanya
    Purini, Suresh
    ACM TRANSACTIONS ON ARCHITECTURE AND CODE OPTIMIZATION, 2022, 19 (03)
  • [25] PROBABILISTIC FINE-GRAINED URBAN FLOW INFERENCE WITH NORMALIZING FLOWS
    Zhong, Ting
    Yu, Haoyang
    Li, Rongfan
    Xu, Xovee
    Luo, Xucheng
    Zhou, Fan
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 3663 - 3667
  • [26] A Hardware Acceleration Platform for AI-Based Inference at the Edge
    Kimon Karras
    Evangelos Pallis
    George Mastorakis
    Yannis Nikoloudakis
    Jordi Mongay Batalla
    Constandinos X. Mavromoustakis
    Evangelos Markakis
    Circuits, Systems, and Signal Processing, 2020, 39 : 1059 - 1070
  • [27] A Hardware Acceleration Platform for AI-Based Inference at the Edge
    Karras, Kimon
    Pallis, Evangelos
    Mastorakis, George
    Nikoloudakis, Yannis
    Batalla, Jordi Mongay
    Mavromoustakis, Constandinos X.
    Markakis, Evangelos
    CIRCUITS SYSTEMS AND SIGNAL PROCESSING, 2020, 39 (02) : 1059 - 1070
  • [28] Edge-Side Fine-Grained Sparse CNN Accelerator With Efficient Dynamic Pruning Scheme
    Wu, Bi
    Yu, Tianyang
    Chen, Ke
    Liu, Weiqiang
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2024, 71 (03) : 1285 - 1298
  • [29] Energy-efficient Oriented Approximate Quantization Scheme for Fine-Grained Sparse Neural Network Acceleration
    Yu, Tianyang
    Wu, Bi
    Chen, Ke
    Yan, Chenggang
    Liu, Weiqiang
    2022 IEEE 40TH INTERNATIONAL CONFERENCE ON COMPUTER DESIGN (ICCD 2022), 2022, : 762 - 769
  • [30] Predictive Exit: Prediction of Fine-Grained Early Exits for Computation- and Energy-Efficient Inference
    Li, Xiangjie
    Lou, Chenfei
    Chen, Yuchi
    Zhu, Zhengping
    Shen, Yingtao
    Ma, Yehan
    Zou, An
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 7, 2023, : 8657 - 8665