Reinforcement Learning From Hierarchical Critics

被引:7
|
作者
Cao, Zehong [1 ]
Lin, Chin-Teng [2 ]
机构
[1] Univ South Australia, STEM, Adelaide, SA 5095, Australia
[2] Univ Technol Sydney, Australian Artificial Intelligence Inst AAII, Sch Comp Sci, Sydney, NSW 2007, Australia
基金
澳大利亚研究理事会;
关键词
Task analysis; Training; Sports; Linear programming; Games; Reinforcement learning; Optimization; Competition; critics; hierarchy; reinforcement learning (RL);
D O I
10.1109/TNNLS.2021.3103642
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this study, we investigate the use of global information to speed up the learning process and increase the cumulative rewards of reinforcement learning (RL) in competition tasks. Within the framework of actor-critic RL, we introduce multiple cooperative critics from two levels of a hierarchy and propose an RL from the hierarchical critics (RLHC) algorithm. In our approach, each agent receives value information from local and global critics regarding a competition task and accesses multiple cooperative critics in a top-down hierarchy. Thus, each agent not only receives low-level details, but also considers coordination from higher levels, thereby obtaining global information to improve the training performance. Then, we test the proposed RLHC algorithm against a benchmark algorithm, that is, proximal policy optimization (PPO), under four experimental scenarios consisting of tennis, soccer, banana collection, and crawler competitions within the Unity environment. The results show that RLHC outperforms the benchmark on these four competitive tasks.
引用
收藏
页码:1066 / 1073
页数:8
相关论文
共 50 条
  • [1] A Reinforcement Learning Method with Implicit Critics from a Bystander
    Hwang, Kao-Shing
    Hsieh, Chi-Wei
    Jiang, Wei-Cheng
    Lin, Jin-Ling
    [J]. ADVANCES IN NEURAL NETWORKS, PT I, 2017, 10261 : 363 - 370
  • [2] Concurrent Hierarchical Reinforcement Learning
    Marthi, Bhaskara
    Russell, Stuart
    Latham, David
    Guestrin, Carlos
    [J]. 19TH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI-05), 2005, : 779 - 785
  • [3] Hierarchical reinforcement learning with OMQ
    Shen, Jing
    Liu, Haibo
    Gu, Guochang
    [J]. PROCEEDINGS OF THE FIFTH IEEE INTERNATIONAL CONFERENCE ON COGNITIVE INFORMATICS, VOLS 1 AND 2, 2006, : 584 - 588
  • [4] Hierarchical Imitation and Reinforcement Learning
    Le, Hoang M.
    Jiang, Nan
    Agarwal, Alekh
    Dudik, Miroslav
    Yue, Yisong
    Daume, Hal, III
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [5] On Efficiency in Hierarchical Reinforcement Learning
    Wen, Zheng
    Precup, Doina
    Ibrahimi, Morteza
    Barreto, Andre
    Van Roy, Benjamin
    Singh, Satinder
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [6] Budgeted Hierarchical Reinforcement Learning
    Leon, Aurelia
    Denoyer, Ludovic
    [J]. 2018 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2018,
  • [7] On Centralized Critics in Multi-Agent Reinforcement Learning
    Lyu, Xueguang
    Baisero, Andrea
    Xiao, Yuchen
    Daley, Brett
    Amato, Christopher
    [J]. Journal of Artificial Intelligence Research, 2023, 77 : 295 - 354
  • [8] On Centralized Critics in Multi-Agent Reinforcement Learning
    Lyu, Xueguang
    Baisero, Andrea
    Xiao, Yuchen
    Daley, Brett
    Amato, Christopher
    [J]. JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2023, 77 : 295 - 354
  • [9] Deep Reinforcement Learning with Hierarchical Structures
    Li, Siyuan
    [J]. PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 4899 - 4900
  • [10] Hierarchical Reinforcement Learning: A Comprehensive Survey
    Pateria, Shubham
    Subagdja, Budhitama
    Tan, Ah-hwee
    Quek, Chai
    [J]. ACM COMPUTING SURVEYS, 2021, 54 (05)