Automated State Feature Learning for Actor-Critic Reinforcement Learning through NEAT

被引:2
|
作者
Peng, Yiming [1 ]
Chen, Gang [1 ]
Holdaway, Scott [1 ]
Mei, Yi [1 ]
Zhang, Mengjie [1 ]
机构
[1] Victoria Univ Wellington, Sch Engn & Comp Sci, Wellington, New Zealand
关键词
NeuroEvolution; NEAT; Actor-Critic; Reinforcement Learning; Feature Extraction; Feature Learning;
D O I
10.1145/3067695.3076035
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Actor-Critic (AC) algorithms are important approaches to solving sophisticated reinforcement learning problems. However, the learning performance of these algorithms rely heavily on good state features that are often designed manually. To address this issue, we propose to adopt an evolutionary approach based on NeuroEvolution of Augmenting Topology (NEAT) to automatically evolve neural networks that directly transform the raw environmental inputs into state features. Following this idea, we have successfully developed a new algorithm called NEAT+AC which combines Regular-gradient Actor-Critic (RAC) with NEAT. It can simultaneously learn suitable state features as well as good policies that are expected to significantly improve the reinforcement learning performance. Preliminary experiments on two benchmark problems confirm that our new algorithm can clearly outperform the baseline algorithm, i.e., NEAT.
引用
收藏
页码:135 / 136
页数:2
相关论文
共 50 条
  • [41] Dynamic Charging Scheme Problem With Actor-Critic Reinforcement Learning
    Yang, Meiyi
    Liu, Nianbo
    Zuo, Lin
    Feng, Yong
    Liu, Minghui
    Gong, Haigang
    Liu, Ming
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2021, 8 (01) : 370 - 380
  • [42] Learning State Representation for Deep Actor-Critic Control
    Munk, Jelle
    Kober, Jens
    Babuska, Robert
    [J]. 2016 IEEE 55TH CONFERENCE ON DECISION AND CONTROL (CDC), 2016, : 4667 - 4673
  • [43] Approximate Optimal Filter Design for Vehicle System through Actor-Critic Reinforcement Learning
    Yin, Yuming
    Li, Shengbo Eben
    Tang, Kaiming
    Cao, Wenhan
    Wu, Wei
    Li, Hongbo
    [J]. AUTOMOTIVE INNOVATION, 2022, 5 (04) : 415 - 426
  • [44] Approximate Optimal Filter Design for Vehicle System through Actor-Critic Reinforcement Learning
    Yuming Yin
    Shengbo Eben Li
    Kaiming Tang
    Wenhan Cao
    Wei Wu
    Hongbo Li
    [J]. Automotive Innovation, 2022, 5 : 415 - 426
  • [45] A Survey of Actor-Critic Reinforcement Learning: Standard and Natural Policy Gradients
    Grondman, Ivo
    Busoniu, Lucian
    Lopes, Gabriel A. D.
    Babuska, Robert
    [J]. IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART C-APPLICATIONS AND REVIEWS, 2012, 42 (06): : 1291 - 1307
  • [46] Bringing Fairness to Actor-Critic Reinforcement Learning for Network Utility Optimization
    Chen, Jingdi
    Wang, Yimeng
    Lan, Tian
    [J]. IEEE CONFERENCE ON COMPUTER COMMUNICATIONS (IEEE INFOCOM 2021), 2021,
  • [47] An extension of Genetic Network Programming with Reinforcement Learning using actor-critic
    Hatakeyama, Hiroyuki
    Mabu, Shingo
    Hirasawa, Kotaro
    Hu, Jinglu
    [J]. 2006 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION, VOLS 1-6, 2006, : 1522 - +
  • [48] Exponential TD Learning: A Risk-Sensitive Actor-Critic Reinforcement Learning Algorithm
    Noorani, Erfaun
    Mavridis, Christos N.
    Baras, John S.
    [J]. 2023 AMERICAN CONTROL CONFERENCE, ACC, 2023, : 4104 - 4109
  • [49] Deep Reinforcement Learning in VizDoom via DQN and Actor-Critic Agents
    Bakhanova, Maria
    Makarov, Ilya
    [J]. ADVANCES IN COMPUTATIONAL INTELLIGENCE, IWANN 2021, PT I, 2021, 12861 : 138 - 150
  • [50] Lexicographic Actor-Critic Deep Reinforcement Learning for Urban Autonomous Driving
    Zhang, Hengrui
    Lin, Youfang
    Han, Sheng
    Lv, Kai
    [J]. IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2023, 72 (04) : 4308 - 4319