An energy efficient EdgeAI autoencoder accelerator for reinforcement learning

被引:9
|
作者
Manjunath N.K. [1 ]
Shiri A. [1 ]
Hosseini M. [1 ]
Prakash B. [1 ]
Waytowich N.R. [2 ]
Mohsenin T. [1 ]
机构
[1] Department of Computer Science and Electrical Engineering, University of Maryland, Baltimore, 21250, MD
[2] U.S. Army Research Laboratory, Aberdeen, MD
关键词
ASIC; autoencoder; autonomous systems; binary neural networks (BNNs); EdgeAI; energy efficiency; FPGA; Reinforcement learning; ternary neural networks (TNNs);
D O I
10.1109/OJCAS.2020.3043737
中图分类号
学科分类号
摘要
In EdgeAI embedded devices that exploit reinforcement learning (RL), it is essential to reduce the number of actions taken by the agent in the real world and minimize the compute-intensive policies learning process. Convolutional autoencoders (AEs) has demonstrated great improvement for speeding up the policy learning time when attached to the RL agent, by compressing the high dimensional input data into a small latent representation for feeding the RL agent. Despite reducing the policy learning time, AE adds a significant computational and memory complexity to the model which contributes to the increase in the total computation and the model size. In this article, we propose a model for speeding up the policy learning process of RL agent with the use of AE neural networks, which engages binary and ternary precision to address the high complexity overhead without deteriorating the policy that an RL agent learns. Binary Neural Networks (BNNs) and Ternary Neural Networks (TNNs) compress weights into 1 and 2 bits representations, which result in significant compression of the model size and memory as well as simplifying multiply-accumulate (MAC) operations. We evaluate the performance of our model in three RL environments including DonkeyCar, Miniworld sidewalk, and Miniworld Object Pickup, which emulate various real-world applications with different levels of complexity. With proper hyperparameter optimization and architecture exploration, TNN models achieve near the same average reward, Peak Signal to Noise Ratio (PSNR) and Mean Squared Error (MSE) performance as the full-precision model while reducing the model size by 10 × compared to full-precision and 3 × compared to BNNs. However, in BNN models the average reward drops up to 12% - 25% compared to the full-precision even after increasing its model size by 4 ×. We designed and implemented a scalable hardware accelerator which is configurable in terms of the number of processing elements (PEs) and memory data width to achieve the best power, performance, and energy efficiency trade-off for EdgeAI embedded devices. The proposed hardware implemented on Artix-7 FPGA dissipates 250 μJ energy while meeting 30 frames per second (FPS) throughput requirements. The hardware is configurable to reach an efficiency of over 1 TOP/J on FPGA implementation. The proposed hardware accelerator is synthesized and placed-and-routed in 14 nm FinFET ASIC technology which brings down the power dissipation to 3.9 μJ and maximum throughput of 1,250 FPS. Compared to the state of the art TNN implementations on the same target platform, our hardware is 5 × and 4.4 × (2.2 × if technology scaled) more energy efficient on FPGA and ASIC, respectively. © 2020 IEEE.
引用
收藏
页码:182 / 195
页数:13
相关论文
共 50 条
  • [21] Deep Adversarial Reinforcement Learning With Noise Compensation by Autoencoder
    Ohashi, Kohei
    Nakanishi, Kosuke
    Sasaki, Wataru
    Yasui, Yuji
    Ishii, Shin
    IEEE ACCESS, 2021, 9 : 143901 - 143912
  • [22] Reinforcement Learning driven Energy Efficient Mobile Communication and Applications
    Asad, Syed Muhammad
    Ozturk, Metin
    Bin Rais, Rao Naveed
    Zoha, Ahmed
    Hussain, Sajjad
    Abbasi, Qammer H.
    Imran, Muhammad Ali
    2019 IEEE 19TH INTERNATIONAL SYMPOSIUM ON SIGNAL PROCESSING AND INFORMATION TECHNOLOGY (ISSPIT 2019), 2019,
  • [23] Energy-Efficient Reinforcement Learning for Motion Planning of AUV
    Wen, Jiayi
    Zhu, Jingwei
    Lin, Yejin
    Zhang, Guichen
    2022 IEEE 9TH INTERNATIONAL CONFERENCE ON UNDERWATER SYSTEM TECHNOLOGY: THEORY AND APPLICATIONS, USYS, 2022,
  • [24] An Energy-Efficient Programmable Mixed Signal Accelerator for Machine Learning Algorithms
    Kang, Mingu
    Srivastava, Prakalp
    Adve, Vikram
    Kim, Nam Sung
    Shanbhag, Naresh R.
    IEEE MICRO, 2019, 39 (05) : 64 - 72
  • [25] An Energy-Efficient Deep Reinforcement Learning FPGA Accelerator for Online Fast Adaptation with Selective Mixed-precision Re-training
    Jo, Wooyoung
    Lee, Juhyoung
    Park, Seunghyun
    Yoo, Hoi-Jun
    IEEE ASIAN SOLID-STATE CIRCUITS CONFERENCE (A-SSCC 2021), 2021,
  • [26] Deep reinforcement learning enabled self-learning control for energy efficient driving
    Qi, Xuewei
    Luo, Yadan
    Wu, Guoyuan
    Boriboonsomsin, Kanok
    Barth, Matthew
    TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2019, 99 : 67 - 81
  • [27] Efficiency of Reinforcement Learning using Polarized Regime by Variational Autoencoder
    Nakai, Masato
    Shibuya, Takeshi
    2022 61ST ANNUAL CONFERENCE OF THE SOCIETY OF INSTRUMENT AND CONTROL ENGINEERS (SICE), 2022, : 128 - 134
  • [28] Reinforcement learning for energy-efficient control of parallel and identical machines
    Loffredo, Alberto
    May, Marvin Carl
    Schaefer, Louis
    Matta, Andrea
    Lanza, Gisela
    CIRP JOURNAL OF MANUFACTURING SCIENCE AND TECHNOLOGY, 2023, 44 : 91 - 103
  • [29] A greedy energy efficient clustering scheme based reinforcement learning for WSNs
    Bourebia, Nour El Houda
    Li, Chunlin
    PEER-TO-PEER NETWORKING AND APPLICATIONS, 2022, 15 (06) : 2572 - 2588
  • [30] Learning Efficient Channels with a Dual Loss Autoencoder
    Granstedt, Jason L.
    Zhou, Weimin
    Anastasio, Mark A.
    MEDICAL IMAGING 2020: IMAGE PERCEPTION, OBSERVER PERFORMANCE, AND TECHNOLOGY ASSESSMENT, 2020, 11316