Selector-Actor-Critic and Tuner-Actor-Critic Algorithms for Reinforcement Learning

被引:0
|
作者
Masadeh, Ala'eddin [1 ]
Wang, Zhengdao [1 ]
Kamal, Ahmed E. [1 ]
机构
[1] ISU, Ames, IA 50011 USA
基金
美国国家科学基金会;
关键词
Reinforcement learning; model-based learning; model-free learning; actor-critic; GAME; GO;
D O I
10.1109/wcsp.2019.8928124
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
This work presents two reinforcement learning (RL) architectures, which mimic rational humans in the way of analyzing the available information and making decisions. The proposed algorithms are called selector-actor-critic (SAC) and tuner-actor-critic (TAC). They are obtained by modifying the well known actor-critic (AC) algorithm. SAC is equipped with an actor, a critic, and a selector. The role of the selector is to determine the most promising action at the current state based on the last estimate from the critic. TAC is model based, and consists of a tuner, a model-learner, an actor, and a critic. After receiving the approximated value of the current state-action pair from the critic and the learned model from the model-learner, the tuner uses the Bellman equation to tune the value of the current state-action pair. Then, this tuned value is used by the actor to optimize the policy. We investigate the performance of the proposed algorithms, and compare with AC algorithm to show the advantages of the proposed algorithms using numerical simulations.
引用
收藏
页数:6
相关论文
共 50 条
  • [21] Natural actor-critic algorithms
    Bhatnagar, Shalabh
    Sutton, Richard S.
    Ghavamzadeh, Mohammad
    Lee, Mark
    [J]. AUTOMATICA, 2009, 45 (11) : 2471 - 2482
  • [22] Distributed Actor-Critic Algorithms for Multiagent Reinforcement Learning Over Directed Graphs
    Dai, Pengcheng
    Yu, Wenwu
    Wang, He
    Baldi, Simone
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (10) : 7210 - 7221
  • [23] Intensive versus non-intensive actor-critic reinforcement learning algorithms
    Wawrzynski, P
    Pacut, A
    [J]. ARTIFICIAL INTELLIGENCE AND SOFT COMPUTING - ICAISC 2004, 2004, 3070 : 934 - 941
  • [24] Reinforcement learning with actor-critic for knowledge graph reasoning
    Linli Zhang
    Dewei Li
    Yugeng Xi
    Shuai Jia
    [J]. Science China Information Sciences, 2020, 63
  • [25] Reinforcement learning with actor-critic for knowledge graph reasoning
    Zhang, Linli
    Li, Dewei
    Xi, Yugeng
    Jia, Shuai
    [J]. SCIENCE CHINA-INFORMATION SCIENCES, 2020, 63 (06)
  • [26] Actor-critic reinforcement learning for bidding in bilateral negotiation
    Arslan, Furkan
    Aydogan, Reyhan
    [J]. TURKISH JOURNAL OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCES, 2022, 30 (05) : 1695 - 1714
  • [27] A Sandpile Model for Reliable Actor-Critic Reinforcement Learning
    Peng, Yiming
    Chen, Gang
    Zhang, Mengjie
    Pang, Shaoning
    [J]. 2017 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2017, : 4014 - 4021
  • [28] Actor-Critic Reinforcement Learning for Tracking Control in Robotics
    Pane, Yudha P.
    Nageshrao, Subramanya P.
    Babuska, Robert
    [J]. 2016 IEEE 55TH CONFERENCE ON DECISION AND CONTROL (CDC), 2016, : 5819 - 5826
  • [29] Visual Navigation with Actor-Critic Deep Reinforcement Learning
    Shao, Kun
    Zhao, Dongbin
    Zhu, Yuanheng
    Zhang, Qichao
    [J]. 2018 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2018,
  • [30] Stochastic Integrated Actor-Critic for Deep Reinforcement Learning
    Zheng, Jiaohao
    Kurt, Mehmet Necip
    Wang, Xiaodong
    [J]. IEEE Transactions on Neural Networks and Learning Systems, 2024, 35 (05) : 6654 - 6666