Accelerating Deep Neuroevolution on Distributed FPGAs for Reinforcement Learning Problems

被引:2
|
作者
Asseman, Alexis [1 ]
Antoine, Nicolas [1 ]
Ozcan, Ahmet S. [1 ]
机构
[1] IBM Almaden Res Ctr, 650 Harry Rd, San Jose, CA 95120 USA
关键词
Genetic algorithm; field programmable gate array; neuroevolution; reinforcement learning; artificial neural network; ENVIRONMENT;
D O I
10.1145/3425500
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Reinforcement learning, augmented by the representational power of deep neural networks, has shown promising results on high-dimensional problems, such as game playing and robotic control. However, the sequential nature of these problems poses a fundamental challenge for computational efficiency. Recently, alternative approaches such as evolutionary strategies and deep neuroevolution demonstrated competitive results with faster training time on distributed CPU cores. Here we report record training times (running at about 1 million frames per second) for Atari 2600 games using deep neuroevolution implemented on distributed FPGAs. Combined hardware implementation of the game console, image preprocessing and the neural network in an optimized pipeline, multiplied with the system level parallelism enabled the acceleration. These results are the first application demonstration on the IBM Neural Computer, which is a custom designed system that consists of 432 Xilinx FPGAs interconnected in a 3D mesh network topology. In addition to high performance, experiments also showed improvement in accuracy for all games compared to the CPU implementation of the same algorithm.
引用
收藏
页数:17
相关论文
共 50 条
  • [21] Accelerating Drugs Discovery with Deep Reinforcement Learning: An Early Approach
    Serrano, Antonio
    Imbernon, Baldomero
    Perez-Sanchez, Horacio
    Cecilia, Jose M.
    Bueno-Crespo, Andres
    Abellan, Jose L.
    47TH INTERNATIONAL CONFERENCE ON PARALLEL PROCESSING (ICPP '18), 2018,
  • [22] Accelerating Sparse Deep Neural Networks on FPGAs
    Huang, Sitao
    Pearson, Carl
    Nagi, Rakesh
    Xiong, Jinjun
    Chen, Deming
    Hwu, Wen-mei
    2019 IEEE HIGH PERFORMANCE EXTREME COMPUTING CONFERENCE (HPEC), 2019,
  • [23] Enhancing Pokemon VGC Player Performance: Intelligent Agents Through Deep Reinforcement Learning and Neuroevolution
    Rodriguez, Gian
    Villanueva, Edwin
    Baldeon, Johan
    HCI IN GAMES, PT I, HCI-GAMES 2024, 2024, 14730 : 275 - 294
  • [24] A Distributed-GPU Deep Reinforcement Learning System for Solving Large Graph Optimization Problems
    Zheng, Weijian
    Wang, Dali
    Song, Fengguang
    ACM TRANSACTIONS ON PARALLEL COMPUTING, 2023, 10 (02)
  • [25] Orchestra: Adaptively Accelerating Distributed Deep Learning in Heterogeneous Environments
    Du, Haizhou
    Huang, Sheng
    Xiang, Qiao
    PROCEEDINGS OF THE 19TH ACM INTERNATIONAL CONFERENCE ON COMPUTING FRONTIERS 2022 (CF 2022), 2022, : 181 - 184
  • [26] Communication Optimization Schemes for Accelerating Distributed Deep Learning Systems
    Lee, Jaehwan
    Choi, Hyeonseong
    Jeong, Hyeonwoo
    Noh, Baekhyeon
    Shin, Ji Sun
    APPLIED SCIENCES-BASEL, 2020, 10 (24): : 1 - 15
  • [27] Surrogate Models for Enhancing the Efficiency of Neuroevolution in Reinforcement Learning
    Stork, Joerg
    Zaefferer, Martin
    Bartz-Beielstein, Thomas
    Eiben, A. E.
    PROCEEDINGS OF THE 2019 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE (GECCO'19), 2019, : 934 - 942
  • [28] Flexible silicon photonic architecture for accelerating distributed deep learning
    Wu, Zhenguo
    Dai, Liang Yuan
    Wang, Yuyang
    Wang, Songli
    Bergman, Keren
    JOURNAL OF OPTICAL COMMUNICATIONS AND NETWORKING, 2024, 16 (02) : A157 - A168
  • [29] Towards accelerating model parallelism in distributed deep learning systems
    Choi, Hyeonseong
    Lee, Byung Hyun
    Chun, Se Young
    Lee, Jaehwan
    PLOS ONE, 2023, 18 (11):
  • [30] DistDL: A Distributed Deep Learning Service Schema with GPU Accelerating
    Wang, Jianzong
    Cheng, Lianglun
    WEB TECHNOLOGIES AND APPLICATIONS (APWEB 2015), 2015, 9313 : 793 - 804