DARL: Distributed Reconfigurable Accelerator for Hyperdimensional Reinforcement Learning

被引:13
|
作者
Chen, Hanning [1 ]
Issa, Mariam [1 ]
Ni, Yang [1 ]
Imani, Mohsen [1 ]
机构
[1] Univ Calif Irvine, Irvine, CA 92717 USA
基金
美国国家科学基金会;
关键词
D O I
10.1145/3508352.3549437
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Reinforcement Learning (RL) is a powerful technology to solve decisionmaking problems such as robotics control. Modern RL algorithms, i.e., Deep Q-Learning, are based on costly and resource hungry deep neural networks. This motivates us to deploy alternative models for powering RL agents on edge devices. Recently, brain-inspired HyperDimensional Computing (HDC) has been introduced as a promising solution for lightweight and efficient machine learning, particularly for classification. In this work, we develop a novel platform capable of real-time hyper dimensional reinforcement learning. Our heterogeneous CPU-FPGA platform, called DARL, maximizes FPGA's computing capabilities by applying hardware optimizations to hyperdimensional computing's critical operations, including hardware -friendly encoder IP, the hypervector chunk fragmentation, and the delayed model update. Aside from hardware innovation, we also extend the platform to basic single agent RL to support multi-agents distributed learning. We evaluate the effectiveness of our approach on OpenAl Gym tasks. Our results show that the FPGA platform provides on average 20x speedup compared to current state-of-the-art hyperdimensional RL methods running on Intel Xeon 6226 CPU. In addition, DARL provides around 4.8x faster and 4.2x higher energy efficiency compared to the state-of-the-art RL accelerator While ensuring a better or comparable quality of learning.
引用
收藏
页数:9
相关论文
共 50 条
  • [41] MSRL: Distributed Reinforcement Learning with Dataflow Fragments
    Zhu, Huanzhou
    Zhao, Bo
    Chen, Gang
    Chen, Weifeng
    Chen, Yijie
    Shi, Liang
    Yang, Yaodong
    Pietzuch, Peter
    Chen, Lei
    PROCEEDINGS OF THE 2023 USENIX ANNUAL TECHNICAL CONFERENCE, 2023, : 977 - 993
  • [42] Distributed deep reinforcement learning for simulation control
    Pawar, Suraj
    Maulik, Romit
    MACHINE LEARNING-SCIENCE AND TECHNOLOGY, 2021, 2 (02):
  • [43] Distributed Multitask Reinforcement Learning with Quadratic Convergence
    Tutunov, Rasul
    Kim, Dongho
    Bou-Ammar, Haitham
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [44] Automatic Curriculum Determination for Deep Reinforcement Learning in Reconfigurable Robots
    Karni, Zohar
    Simhon, Or
    Zarrouk, David
    Berman, Sigal
    IEEE ACCESS, 2024, 12 : 78342 - 78353
  • [45] A Deep Reinforcement Learning Approach for Autonomous Reconfigurable Intelligent Surfaces
    Choi, Hyuckjin
    Nguyen, Ly V.
    Choi, Junil
    Swindlehurst, A. Lee
    2024 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS WORKSHOPS, ICC WORKSHOPS 2024, 2024, : 208 - 213
  • [46] A Survey on Reinforcement Learning for Reconfigurable Intelligent Surfaces in Wireless Communications
    Puspitasari, Annisa Anggun
    Lee, Byung Moo
    SENSORS, 2023, 23 (05)
  • [47] Optimizing Discharge Efficiency of Reconfigurable Battery With Deep Reinforcement Learning
    Jeon, Seunghyeok
    Kim, Jiwon
    Ahn, Junick
    Cha, Hojung
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2020, 39 (11) : 3893 - 3905
  • [48] Active Balancing of Reconfigurable Batteries Using Reinforcement Learning Algorithms
    Jiang, Bowen
    Tang, Junfei
    Liu, Yujing
    Boscaglia, Luca
    2023 IEEE TRANSPORTATION ELECTRIFICATION CONFERENCE & EXPO, ITEC, 2023,
  • [49] Symbolic Representation and Learning With Hyperdimensional Computing
    Mitrokhin, Anton
    Sutor, Peter
    Summers-Stay, Douglas
    Fermueller, Cornelia
    Aloimonos, Yiannis
    FRONTIERS IN ROBOTICS AND AI, 2020, 7
  • [50] Primal-Dual Algorithm for Distributed Reinforcement Learning: Distributed GTD
    Lee, Donghwan
    Yoon, Hyungjin
    Hovakimyan, Naira
    2018 IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2018, : 1967 - 1972