An ARM-based Q-learning algorithm

被引:0
|
作者
Hsu, Yuan-Pao [1 ]
Hwang, Kao-Shing [2 ]
Lin, Hsin-Yi [2 ]
机构
[1] Natl Formosa Univ, Dept Comp Sci & Informat Engn, Yunlin 632, Taiwan
[2] Natl Chung Cheng Univ, Dept Elect Engn, Chiayi 711, Taiwan
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This article presents an algorithm that combines a FAST-based algorithm (Flexible Adaptable-Size Topology), called ARM, and Q-learning algorithm. The ARM is a self organizing architecture. Dynamically adjusting the size of sensitivity regions of each neuron and adaptively pruning one of the redundant neurons, the ARM can preserve resources (available neurons) to accommodate more categories. The Q-learning is a dynamic programming-based reinforcement learning method, in which the learned action-value function, Q, directly approximates Q*, the optimal action-value function, independent of the policy being followed. In the proposed method, the ARM acts as a cluster to categorize input vectors from the outside world. Clustered results are then sent to the Q-learning architecture in order that it learns to present the best actions to the outside world. The effect of the algorithm is shown through computer simulations of the well-known control of balancing an inverted pendulum on a cart.
引用
收藏
页码:11 / +
页数:2
相关论文
共 50 条
  • [1] Backward Q-learning: The combination of Sarsa algorithm and Q-learning
    Wang, Yin-Hao
    Li, Tzuu-Hseng S.
    Lin, Chih-Jui
    [J]. ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2013, 26 (09) : 2184 - 2193
  • [2] Synergism of Firefly Algorithm and Q-Learning for Robot Arm Path Planning
    Sadhu, Arup Kumar
    Konar, Amit
    Bhattacharjee, Tanuka
    Das, Swagatam
    [J]. SWARM AND EVOLUTIONARY COMPUTATION, 2018, 43 : 50 - 68
  • [3] A Task Scheduling Algorithm Based on Q-Learning for WSNs
    Zhang, Benhong
    Wu, Wensheng
    Bi, Xiang
    Wang, Yiming
    [J]. COMMUNICATIONS AND NETWORKING, CHINACOM 2018, 2019, 262 : 521 - 530
  • [4] Power Control Algorithm Based on Q-Learning in Femtocell
    Li Yun
    Tang Ying
    Liu Hanxiao
    [J]. JOURNAL OF ELECTRONICS & INFORMATION TECHNOLOGY, 2019, 41 (11) : 2557 - 2564
  • [5] Q-Learning Algorithm Based on Incremental RBF Network
    Hu, Yanming
    Li, Decai
    He, Yuqing
    Han, Jianda
    [J]. Jiqiren/Robot, 2019, 41 (05): : 562 - 573
  • [6] Coherent beam combination based on Q-learning algorithm
    Zhang, Xi
    Li, Pingxue
    Zhu, Yunchen
    Li, Chunyong
    Yao, Chuanfei
    Wang, Luo
    Dong, Xueyan
    Li, Shun
    [J]. OPTICS COMMUNICATIONS, 2021, 490
  • [7] A new Q-learning algorithm based on the Metropolis criterion
    Guo, MZ
    Liu, Y
    Malec, J
    [J]. IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 2004, 34 (05): : 2140 - 2143
  • [8] Adaptive PID controller based on Q-learning algorithm
    Shi, Qian
    Lam, Hak-Keung
    Xiao, Bo
    Tsai, Shun-Hung
    [J]. CAAI TRANSACTIONS ON INTELLIGENCE TECHNOLOGY, 2018, 3 (04) : 235 - 244
  • [9] Hexagon-based Q-learning algorithm and applications
    Yang, Hyun-Chang
    Kim, Ho-Duck
    Yoon, Han-Ul
    Jang, In-Hun
    Sim, Kwee-Bo
    [J]. INTERNATIONAL JOURNAL OF CONTROL AUTOMATION AND SYSTEMS, 2007, 5 (05) : 570 - 576
  • [10] Ramp Metering Control Based on the Q-Learning Algorithm
    Ivanjko, Edouard
    Necoska, Daniela Koltovska
    Greguric, Martin
    Vujic, Miroslav
    Jurkovic, Goran
    Mandzuka, Sadko
    [J]. CYBERNETICS AND INFORMATION TECHNOLOGIES, 2015, 15 (05) : 88 - 97