An Energy-Efficient Hardware Accelerator for Hierarchical Deep Reinforcement Learning

被引:4
|
作者
Shiri, Aidin [1 ]
Prakash, Bharat [1 ]
Mazumder, Arnab Neelim [1 ]
Waytowich, Nicholas R. [2 ]
Oates, Tim [1 ]
Mohsenin, Tinoosh [1 ]
机构
[1] Univ Maryland Baltimore Cty, Dept Comp Sci & Elect Engn, Baltimore, MD 21228 USA
[2] US Army Res Lab, Aberdeen Proving Ground, MD USA
关键词
Reinforcement Learning; Energy Efficient Hardware; FPGA; ASIC;
D O I
10.1109/AICAS51828.2021.9458548
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Reinforcement Learning (RL) has shown great performance in solving sequential decision-making and control in dynamic environments problems. Despite its achievements, training Deep Neural Network (DNN) based RL is expensive in terms of time and power because of the large number of episodes required to train agents with high dimensional image representations. At the deployment also, the massive energy footprint of deep neural networks can be a major drawback. Embedded devices as the main deployment platform, are intrinsically resource-constrained and deploying DNN on them is challenging. Consequently, reducing the number of actions taken by the RL agent to learn desired policy, along with the development of efficient hardware architectures for RL is crucial. In this paper, we propose a novel hardware architecture for RL agents based on the learning hierarchical policies method. We show that hierarchical learning with several levels of control improves RL agents training efficiency and the agent converges faster compared to a none hierarchical model and therefore using less power. This is especially true as the environment becomes more complex with multiple objective sub-goals. Our method is important for efficient learning of policies for RL agent, especially when the target platform is a resource constraint embedded device. By performing a systematic neural network architecture search and hardware design space exploration, we implemented an energy-efficient scalable hardware accelerator for the hierarchical RL. Hardware factors of merit such as the latency, throughput, and energy consumption of the accelerator are evaluated with the various processing elements, and model parameters. The most energy-efficient configuration achieves 139 fps throughput with 5.8 mJ energy consumption per classification on Xilinx Artix-7 FPGA. Compared to similar works our design shows up to 3x better energy efficiency.
引用
收藏
页数:4
相关论文
共 50 条
  • [1] E2HRL: An Energy-efficient Hardware Accelerator for Hierarchical Deep Reinforcement Learning
    Shiri, Aidin
    Kallakuri, Uttej
    Rashid, Hasib-Al
    Prakash, Bharat
    Waytowich, Nicholas R.
    Oates, Tim
    Mohsenin, Tinoosh
    [J]. ACM TRANSACTIONS ON DESIGN AUTOMATION OF ELECTRONIC SYSTEMS, 2022, 27 (05)
  • [2] Energy-Efficient Deep Reinforcement Learning Accelerator Designs for Mobile Autonomous Systems
    Lee, Juhyoung
    Kim, Changhyeon
    Han, Donghyeon
    Kim, Sangyeob
    Kim, Sangjin
    Yoo, Hoi-Jun
    [J]. 2021 IEEE 3RD INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE CIRCUITS AND SYSTEMS (AICAS), 2021,
  • [3] An Energy-Efficient Deep Reinforcement Learning Accelerator With Transposable PE Array and Experience Compression
    Kim, Changhyeon
    Kang, Sanghoon
    Choi, Sungpill
    Shin, Dongjoo
    Kim, Youngwoo
    Yoo, Hoi-Jun
    [J]. IEEE SOLID-STATE CIRCUITS LETTERS, 2019, 2 (11): : 228 - 231
  • [4] Joint Edge Association and Aggregation Frequency for Energy-Efficient Hierarchical Federated Learning by Deep Reinforcement Learning
    Ren, Yijing
    Wu, Changxiang
    So, Daniel K. C.
    [J]. ICC 2023-IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS, 2023, : 3639 - 3645
  • [5] Hierarchical Multi-Agent Deep Reinforcement Learning for Energy-Efficient Hybrid Computation Offloading
    Zhou, Hang
    Long, Yusi
    Gong, Shimin
    Zhu, Kun
    Hoang, Dinh Thai
    Niyato, Dusit
    [J]. IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2023, 72 (01) : 986 - 1001
  • [6] Energy-efficient VM scheduling based on deep reinforcement learning
    Wang, Bin
    Liu, Fagui
    Lin, Weiwei
    [J]. FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2021, 125 : 616 - 628
  • [7] Energy-Efficient IoT Sensor Calibration With Deep Reinforcement Learning
    Ashiquzzaman, Akm
    Lee, Hyunmin
    Um, Tai-Won
    Kim, Jinsul
    [J]. IEEE ACCESS, 2020, 8 : 97045 - 97055
  • [8] Hierarchical Reinforcement Learning for RIS-Assisted Energy-Efficient RAN
    Zhou, Hao
    Kong, Long
    Elsayed, Medhat
    Bavand, Majid
    Gaigalas, Raimundas
    Furr, Steve
    Erol-Kantarci, Melike
    [J]. 2022 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM 2022), 2022, : 3326 - 3331
  • [9] Reinforcement co-Learning of Deep and Spiking Neural Networks for Energy-Efficient Mapless Navigation with Neuromorphic Hardware
    Tang, Guangzhi
    Kumar, Neelesh
    Michmizos, Konstantinos P.
    [J]. 2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 6090 - 6097
  • [10] Energy-Efficient Ultra-Dense Network With Deep Reinforcement Learning
    Ju, Hyungyu
    Kim, Seungnyun
    Kim, Youngjoon
    Shim, Byonghyo
    [J]. IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2022, 21 (08) : 6539 - 6552