An energy efficient EdgeAI autoencoder accelerator for reinforcement learning

被引:9
|
作者
Manjunath N.K. [1 ]
Shiri A. [1 ]
Hosseini M. [1 ]
Prakash B. [1 ]
Waytowich N.R. [2 ]
Mohsenin T. [1 ]
机构
[1] Department of Computer Science and Electrical Engineering, University of Maryland, Baltimore, 21250, MD
[2] U.S. Army Research Laboratory, Aberdeen, MD
关键词
ASIC; autoencoder; autonomous systems; binary neural networks (BNNs); EdgeAI; energy efficiency; FPGA; Reinforcement learning; ternary neural networks (TNNs);
D O I
10.1109/OJCAS.2020.3043737
中图分类号
学科分类号
摘要
In EdgeAI embedded devices that exploit reinforcement learning (RL), it is essential to reduce the number of actions taken by the agent in the real world and minimize the compute-intensive policies learning process. Convolutional autoencoders (AEs) has demonstrated great improvement for speeding up the policy learning time when attached to the RL agent, by compressing the high dimensional input data into a small latent representation for feeding the RL agent. Despite reducing the policy learning time, AE adds a significant computational and memory complexity to the model which contributes to the increase in the total computation and the model size. In this article, we propose a model for speeding up the policy learning process of RL agent with the use of AE neural networks, which engages binary and ternary precision to address the high complexity overhead without deteriorating the policy that an RL agent learns. Binary Neural Networks (BNNs) and Ternary Neural Networks (TNNs) compress weights into 1 and 2 bits representations, which result in significant compression of the model size and memory as well as simplifying multiply-accumulate (MAC) operations. We evaluate the performance of our model in three RL environments including DonkeyCar, Miniworld sidewalk, and Miniworld Object Pickup, which emulate various real-world applications with different levels of complexity. With proper hyperparameter optimization and architecture exploration, TNN models achieve near the same average reward, Peak Signal to Noise Ratio (PSNR) and Mean Squared Error (MSE) performance as the full-precision model while reducing the model size by 10 × compared to full-precision and 3 × compared to BNNs. However, in BNN models the average reward drops up to 12% - 25% compared to the full-precision even after increasing its model size by 4 ×. We designed and implemented a scalable hardware accelerator which is configurable in terms of the number of processing elements (PEs) and memory data width to achieve the best power, performance, and energy efficiency trade-off for EdgeAI embedded devices. The proposed hardware implemented on Artix-7 FPGA dissipates 250 μJ energy while meeting 30 frames per second (FPS) throughput requirements. The hardware is configurable to reach an efficiency of over 1 TOP/J on FPGA implementation. The proposed hardware accelerator is synthesized and placed-and-routed in 14 nm FinFET ASIC technology which brings down the power dissipation to 3.9 μJ and maximum throughput of 1,250 FPS. Compared to the state of the art TNN implementations on the same target platform, our hardware is 5 × and 4.4 × (2.2 × if technology scaled) more energy efficient on FPGA and ASIC, respectively. © 2020 IEEE.
引用
收藏
页码:182 / 195
页数:13
相关论文
共 50 条
  • [1] An Energy-Efficient Hardware Accelerator for Hierarchical Deep Reinforcement Learning
    Shiri, Aidin
    Prakash, Bharat
    Mazumder, Arnab Neelim
    Waytowich, Nicholas R.
    Oates, Tim
    Mohsenin, Tinoosh
    2021 IEEE 3RD INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE CIRCUITS AND SYSTEMS (AICAS), 2021,
  • [2] Energy-Efficient Deep Reinforcement Learning Accelerator Designs for Mobile Autonomous Systems
    Lee, Juhyoung
    Kim, Changhyeon
    Han, Donghyeon
    Kim, Sangyeob
    Kim, Sangjin
    Yoo, Hoi-Jun
    2021 IEEE 3RD INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE CIRCUITS AND SYSTEMS (AICAS), 2021,
  • [3] Neutrons Sensitivity of Deep Reinforcement Learning Policies on EdgeAI Accelerators
    Bodmann, Pablo R.
    Saveriano, Matteo
    Kritikakou, Angeliki
    Rech, Paolo
    IEEE TRANSACTIONS ON NUCLEAR SCIENCE, 2024, 71 (08) : 1480 - 1486
  • [4] An Energy-Efficient Deep Reinforcement Learning Accelerator With Transposable PE Array and Experience Compression
    Kim, Changhyeon
    Kang, Sanghoon
    Choi, Sungpill
    Shin, Dongjoo
    Kim, Youngwoo
    Yoo, Hoi-Jun
    IEEE SOLID-STATE CIRCUITS LETTERS, 2019, 2 (11): : 228 - 231
  • [5] Service Chaining Offloading Decision in the EdgeAI: A Deep Reinforcement Learning Approach
    Lee, Minkyung
    Hong, Choong Seon
    APNOMS 2020: 2020 21ST ASIA-PACIFIC NETWORK OPERATIONS AND MANAGEMENT SYMPOSIUM (APNOMS), 2020, : 393 - 396
  • [6] Efficient beam commissioning in HIPI accelerator based on reinforcement learning
    Su, Chunguang
    Wang, Zhijun
    Chen, Xiaolong
    Jia, Yongzhi
    Qi, Xin
    Wang, Wangsheng
    Sun, Kunxiang
    Du, Yu
    Wang, Tielong
    Chu, Yimeng
    NUCLEAR INSTRUMENTS & METHODS IN PHYSICS RESEARCH SECTION A-ACCELERATORS SPECTROMETERS DETECTORS AND ASSOCIATED EQUIPMENT, 2025, 1072
  • [7] Sample-efficient reinforcement learning for CERN accelerator control
    Kain, Verena
    Hirlander, Simon
    Goddard, Brennan
    Velotti, Francesco Maria
    Porta, Giovanni Zevi Della
    Bruchon, Niky
    Valentino, Gianluca
    PHYSICAL REVIEW ACCELERATORS AND BEAMS, 2020, 23 (12)
  • [8] E2HRL: An Energy-efficient Hardware Accelerator for Hierarchical Deep Reinforcement Learning
    Shiri, Aidin
    Kallakuri, Uttej
    Rashid, Hasib-Al
    Prakash, Bharat
    Waytowich, Nicholas R.
    Oates, Tim
    Mohsenin, Tinoosh
    ACM TRANSACTIONS ON DESIGN AUTOMATION OF ELECTRONIC SYSTEMS, 2022, 27 (05)
  • [9] MEMTONIC: A Neurornorphic Accelerator for Energy Efficient Deep Learning
    Dang, Dharanidhar
    Taheri, Sahar
    Lin, Bill
    Sahoo, Debashis
    PROCEEDINGS OF THE 2020 57TH ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2020,
  • [10] Efficient Accelerator/Network Co-Search With Circular Greedy Reinforcement Learning
    Liu, Zejian
    Li, Gang
    Cheng, Jian
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2023, 70 (07) : 2615 - 2619