Network Parameter Setting for Reinforcement Learning Approaches Using Neural Networks

被引:2
|
作者
Yamada, Kazuaki [1 ]
机构
[1] Toyo Univ, Fac Sci & Engn, Dept Mech Engn, 2100 Kujirai, Kawagoe, Saitama 3508585, Japan
关键词
reinforcement learning; artificial neural networks; autonomous mobile robot;
D O I
10.20965/jaciii.2011.p0822
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Reinforcement learning approaches are attracting attention as a technique for constructing a trial-and-error mapping function between sensors and motors of an autonomous mobile robot. Conventional reinforcement learning approaches use a look-up table to express the mapping function between grid state and grid action spaces. The grid size greatly adversely affects the learning performance of reinforcement learning algorithms. To avoid this, researchers have proposed reinforcement learning algorithms using neural networks to express the mapping function between continuous state space and action. A designer, however, must set the number of middle neurons and initial values of weight parameters appropriately to improve the approximate accuracy of neural networks. This paper proposes a new method that automatically sets the number of middle neurons and initial values of weight parameters based on the dimension number of the sensor space. The feasibility of proposed method is demonstrated using an autonomous mobile robot navigation problem and is evaluated by comparing it with two types of Q-learning as follows: Q-learning using RBF networks and Q-learning using neural networks whose parameters are set by a designer.
引用
收藏
页码:822 / 830
页数:9
相关论文
共 50 条
  • [21] Application of Reinforcement Learning Algorithms for the Adaptive Computation of the Smoothing Parameter for Probabilistic Neural Network
    Kusy, Maciej
    Zajdel, Roman
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2015, 26 (09) : 2163 - 2175
  • [22] Neural Network Ensembles in Reinforcement Learning
    Stefan Faußer
    Friedhelm Schwenker
    Neural Processing Letters, 2015, 41 : 55 - 69
  • [23] Neural Network Ensembles in Reinforcement Learning
    Fausser, Stefan
    Schwenker, Friedhelm
    NEURAL PROCESSING LETTERS, 2015, 41 (01) : 55 - 69
  • [24] Deep reinforcement learning guided graph neural networks for brain network analysis
    Zhao, Xusheng
    Wu, Jia
    Peng, Hao
    Beheshti, Amin
    Monaghan, Jessica J. M.
    McAlpine, David
    Hernandez-Perez, Heivet
    Dras, Mark
    Dai, Qiong
    Li, Yangyang
    Yu, Philip S.
    He, Lifang
    NEURAL NETWORKS, 2022, 154 : 56 - 67
  • [25] Training Spiking Neural Networks Using Combined Learning Approaches
    Elbrecht, Daniel
    Parsa, Maryam
    Kulkarni, Shruti R.
    Mitchell, J. Parker
    Schuman, Catherine D.
    2020 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (SSCI), 2020, : 1995 - 2001
  • [26] Training Neural Networks to Play Backgammon Variants Using Reinforcement Learning
    Papahristou, Nikolaos
    Refanidis, Ioannis
    APPLICATIONS OF EVOLUTIONARY COMPUTATION, PT I, 2011, 6624 : 113 - 122
  • [27] Using Neuroevolved Binary Neural Networks to solve reinforcement learning environments
    Valencia, Raul
    Sham, Chiu-Wing
    Sinnen, Oliver
    2019 IEEE ASIA PACIFIC CONFERENCE ON CIRCUITS AND SYSTEMS (APCCAS 2019), 2019, : 301 - 304
  • [28] Reinforcement Learning using Physics Inspired Graph Convolutional Neural Networks
    Wu, Tong
    Scaglione, Anna
    Arnold, Daniel
    2022 58TH ANNUAL ALLERTON CONFERENCE ON COMMUNICATION, CONTROL, AND COMPUTING (ALLERTON), 2022,
  • [29] Reinforcement learning using swarm intelligence-trained neural networks
    Conforth, M.
    Meng, Y.
    JOURNAL OF EXPERIMENTAL & THEORETICAL ARTIFICIAL INTELLIGENCE, 2010, 22 (03) : 197 - 218
  • [30] Assigning discounts in a marketing campaign by using reinforcement learning and neural networks
    Gomez-Perez, Gabriel
    Martin-Guerrero, Jose D.
    Soria-Olivas, Emilio
    Balaguer-Ballester, Emili
    Palomares, Alberto
    Casariego, Nicolas
    EXPERT SYSTEMS WITH APPLICATIONS, 2009, 36 (04) : 8022 - 8031