DARL: Distributed Reconfigurable Accelerator for Hyperdimensional Reinforcement Learning

被引:13
|
作者
Chen, Hanning [1 ]
Issa, Mariam [1 ]
Ni, Yang [1 ]
Imani, Mohsen [1 ]
机构
[1] Univ Calif Irvine, Irvine, CA 92717 USA
基金
美国国家科学基金会;
关键词
D O I
10.1145/3508352.3549437
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Reinforcement Learning (RL) is a powerful technology to solve decisionmaking problems such as robotics control. Modern RL algorithms, i.e., Deep Q-Learning, are based on costly and resource hungry deep neural networks. This motivates us to deploy alternative models for powering RL agents on edge devices. Recently, brain-inspired HyperDimensional Computing (HDC) has been introduced as a promising solution for lightweight and efficient machine learning, particularly for classification. In this work, we develop a novel platform capable of real-time hyper dimensional reinforcement learning. Our heterogeneous CPU-FPGA platform, called DARL, maximizes FPGA's computing capabilities by applying hardware optimizations to hyperdimensional computing's critical operations, including hardware -friendly encoder IP, the hypervector chunk fragmentation, and the delayed model update. Aside from hardware innovation, we also extend the platform to basic single agent RL to support multi-agents distributed learning. We evaluate the effectiveness of our approach on OpenAl Gym tasks. Our results show that the FPGA platform provides on average 20x speedup compared to current state-of-the-art hyperdimensional RL methods running on Intel Xeon 6226 CPU. In addition, DARL provides around 4.8x faster and 4.2x higher energy efficiency compared to the state-of-the-art RL accelerator While ensuring a better or comparable quality of learning.
引用
收藏
页数:9
相关论文
共 50 条
  • [31] Distributed Reinforcement Learning with ADMM-RL
    Graf, Peter
    Annoni, Jennifer
    Bay, Christopher
    Biagioni, Dave
    Sigler, Devon
    Lunacek, Monte
    Jones, Wesley
    2019 AMERICAN CONTROL CONFERENCE (ACC), 2019, : 4159 - 4166
  • [32] A Distributed Framework for Deep Reinforcement Learning by Consensus
    Liu, Bo
    Zhu, Shuang
    Sun, Peng
    Huang, Qisheng
    Ding, Zhenglao
    2023 35TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2023, : 269 - 274
  • [33] Distributed Spectrum Management based on Reinforcement Learning
    Bernardo, Francisco
    Agusti, Ramon
    Perez-Romero, Jordi
    Sallent, Oriol
    2009 4TH INTERNATIONAL CONFERENCE ON COGNITIVE RADIO ORIENTED WIRELESS NETWORKS AND COMMUNICATIONS, 2009, : 306 - 311
  • [34] Efficient Distributed Reinforcement Learning through Agreement
    Varshavskaya, Paulina
    Kaelbling, Leslie Pack
    Rus, Daniela
    DISTRIBUTED AUTONOMOUS ROBOTIC SYSTEMS 8, 2009, : 367 - 378
  • [35] Distributed Reinforcement Learning for Robot Teams: a Review
    Yutong Wang
    Mehul Damani
    Pamela Wang
    Yuhong Cao
    Guillaume Sartoretti
    Current Robotics Reports, 2022, 3 (4): : 239 - 257
  • [36] Train Throughput Analysis of Distributed Reinforcement Learning
    Jang, Sooyoung
    Park, Noh-Sam
    11TH INTERNATIONAL CONFERENCE ON ICT CONVERGENCE: DATA, NETWORK, AND AI IN THE AGE OF UNTACT (ICTC 2020), 2020, : 1189 - 1192
  • [37] Distributed Reinforcement Learning in Emergency Response Simulation
    Lopez, Cesar
    Marti, Jose R.
    Sarkaria, Sarbjit
    IEEE ACCESS, 2018, 6 : 67261 - 67276
  • [38] Distributed Deep Reinforcement Learning using TensorFlow
    Rao, P. Ajay
    Kumar, Navaneesh B.
    Cadabam, Siddharth
    Praveena, T.
    2017 INTERNATIONAL CONFERENCE ON CURRENT TRENDS IN COMPUTER, ELECTRICAL, ELECTRONICS AND COMMUNICATION (CTCEEC), 2017, : 171 - 174
  • [39] Distributed Emergent Agreements with Deep Reinforcement Learning
    Schmid, Kyrill
    Mueller, Robert
    Belzner, Lenz
    Tochtermann, Johannes
    Linhoff-Popien, Claudia
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [40] Metaoptimization on a Distributed System for Deep Reinforcement Learning
    Heinrich, Greg
    Frosio, Iuri
    PROCEEDINGS OF 2019 5TH IEEE/ACM WORKSHOP ON MACHINE LEARNING IN HIGH PERFORMANCE COMPUTING ENVIRONMENTS (MLHPC 2019), 2019, : 19 - 30