Multitask Augmented Random Search in deep reinforcement learning

被引:1
|
作者
Thanh, Le Tien [1 ]
Thang, Ta Bao [2 ]
Van Cuong, Le [2 ]
Binh, Huynh Thi Thanh [2 ]
机构
[1] Grad Univ Adv Studies, Dept Informat, SOKENDAI, Tokyo, Japan
[2] Hanoi Univ Sci & Technol, Sch Informat & Commun Technol, Hanoi, Vietnam
关键词
Multifactorial evolutionary algorithm; Deep reinforcement learning; Augmented random search;
D O I
10.1016/j.asoc.2024.111605
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Reinforcement Learning (RL) has gained significant popularity in recent years for its ability to solve complex control problems. However, most existing RL algorithms are designed to train policies for each environment in isolation, limiting their applicability to real -world scenarios with many related environments. Recently, many multitask optimization algorithms have been proposed and successfully applied to a wide range of optimization problems. However, existing studies mainly focus on solving multiple continuous function benchmarks and overlook the potential of tailoring towards RL. In this paper, we propose a simple multitask optimization algorithm called Multitask Augmented Random Search (MARS) that trains multiple RL agents together and exploits the performance surplus from highly correlated tasks. MARS is a modification of the simple random search Augmented Random Search (ARS) algorithm, which has been shown to outperform complicated methods in solving continuous control MuJoCo environments such as Soft Actor -Critic (SAC), Proximal Policy Optimization (PPO), and Trusted Region Policy Optimization (TRPO). The experimental results also demonstrate that our proposed algorithm is more consistent in solving different instances of MuJoCo benchmark than ARS, Multifactorial Evolution Algorithm (MFEA), and Adaptive MFEA RL (AMFEARL) within the same number of training episodes.
引用
收藏
页数:10
相关论文
共 50 条
  • [11] Diversity-augmented intrinsic motivation for deep reinforcement learning
    Dai, Tianhong
    Du, Yali
    Fang, Meng
    Bharath, Anil Anthony
    [J]. NEUROCOMPUTING, 2022, 468 : 396 - 406
  • [12] Trust-Augmented Deep Reinforcement Learning for Federated Learning Client Selection
    Rjoub, Gaith
    Wahab, Omar Abdel
    Bentahar, Jamal
    Cohen, Robin
    Bataineh, Ahmed Saleh
    [J]. INFORMATION SYSTEMS FRONTIERS, 2024, 26 (04) : 1261 - 1278
  • [13] Collaborative Deep Reinforcement Learning for Joint Object Search
    Kong, Xiangyu
    Xin, Bo
    Wang, Yizhou
    Hua, Gang
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 7072 - 7081
  • [14] Efficient Novelty Search Through Deep Reinforcement Learning
    Shi, Longxiang
    Li, Shijian
    Zheng, Qian
    Yao, Min
    Pan, Gang
    [J]. IEEE ACCESS, 2020, 8 : 128809 - 128818
  • [15] Provable Benefit of Multitask Representation Learning in Reinforcement Learning
    Cheng, Yuan
    Feng, Songtao
    Yang, Jing
    Zhang, Hong
    Liang, Yingbin
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [16] Learning to Control Random Boolean Networks: A Deep Reinforcement Learning Approach
    Papagiannis, Georgios
    Moschoyiannis, Sotiris
    [J]. COMPLEX NETWORKS AND THEIR APPLICATIONS VIII, VOL 1, 2020, 881 : 721 - 734
  • [17] Scalable Multitask Policy Gradient Reinforcement Learning
    El Bsat, Salam
    Ammar, Haitham Bou
    Taylor, Matthew E.
    [J]. THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 1847 - 1853
  • [18] Evolutionary computation on multitask reinforcement learning problems
    Handa, Hisashi
    [J]. 2007 IEEE INTERNATIONAL CONFERENCE ON NETWORKING, SENSING, AND CONTROL, VOLS 1 AND 2, 2007, : 685 - 688
  • [19] Evolutionary Architecture Search For Deep Multitask Networks
    Liang, Jason
    Meyerson, Elliot
    Miikkulainen, Risto
    [J]. GECCO'18: PROCEEDINGS OF THE 2018 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE, 2018, : 466 - 473
  • [20] Distributed Multitask Reinforcement Learning with Quadratic Convergence
    Tutunov, Rasul
    Kim, Dongho
    Bou-Ammar, Haitham
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31