Competitive reinforcement learning in continuous control tasks

被引:0
|
作者
Abramson, M [1 ]
Pachowicz, P [1 ]
Wechsler, H [1 ]
机构
[1] George Mason Univ, Fairfax, VA 22030 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper describes a novel hybrid reinforcement learning algorithm, Sarsa Learning Vector Quantization (SLVQ), that leaves the reinforcement part intact but employs a more effective representation of the policy function using,a piecewise constant function based upon "policy prototypes." The prototypes correspond to the pattern classes induced by the Voronoi tessellation generated by self-organizing methods like Learning Vector Quantization (LVQ). The determination of the optimal policy function can be now viewed as a pattern recognition problem in the sense that the assignment of an action to a point in the phase space is similar to the assignment of a pattern class to a point in phase space. The distributed LVQ representation of the policy function automatically generates a piecewise constant tessellation of the state space and yields in a major simplification of the learning task relative to the standard reinforcement learning algorithms for whom a discontinuous table look function, has to be learned. The feasibility and comparative advantages of the new algorithm is shown on the cart centering and mountain car problems, two control problems of increased difficulty.
引用
收藏
页码:1909 / 1914
页数:6
相关论文
共 50 条
  • [1] Experiments of conditioned reinforcement learning in continuous space control tasks
    Fernandez-Gauna, Borja
    Osa, Juan Luis
    Grana, Manuel
    [J]. NEUROCOMPUTING, 2018, 271 : 38 - 47
  • [2] Two Steps Reinforcement Learning in Continuous Reinforcement Learning Tasks
    Lopez-Bueno, Ivan
    Garcia, Javier
    Fernandez, Fernando
    [J]. BIO-INSPIRED SYSTEMS: COMPUTATIONAL AND AMBIENT INTELLIGENCE, PT 1, 2009, 5517 : 577 - 584
  • [3] Robust Reinforcement Learning in Continuous Control Tasks with Uncertainty Set Regularization
    Zhang, Yuan
    Wang, Jianhong
    Boedecker, Joschka
    [J]. CONFERENCE ON ROBOT LEARNING, VOL 229, 2023, 229
  • [4] The Strategy for Lane-keeping Vehicle Tasks based on Deep Reinforcement Learning Continuous Control
    Li, Qianxi
    Fei, Rong
    [J]. PROCEEDINGS OF 2024 INTERNATIONAL CONFERENCE ON MACHINE INTELLIGENCE AND DIGITAL APPLICATIONS, MIDA2024, 2024, : 724 - 730
  • [5] Reinforcement Learning and Robust Control for Robot Compliance Tasks
    Cheng-Peng Kuan
    Kuu-young Young
    [J]. Journal of Intelligent and Robotic Systems, 1998, 23 : 165 - 182
  • [6] Reinforcement learning and robust control for robot compliance tasks
    Kuan, CP
    Young, KY
    [J]. JOURNAL OF INTELLIGENT & ROBOTIC SYSTEMS, 1998, 23 (2-4) : 165 - 182
  • [7] Kernel Reinforcement Learning-Assisted Adaptive Decoder Facilitates Stable and Continuous Brain Control Tasks
    Zhang, Xiang
    Chen, Shuhang
    Wang, Yiwen
    [J]. IEEE TRANSACTIONS ON NEURAL SYSTEMS AND REHABILITATION ENGINEERING, 2023, 31 : 4125 - 4134
  • [8] Task-oriented reinforcement learning for continuous tasks in dynamic environment
    Kamal, MAS
    Murata, J
    Hirasawa, K
    [J]. SICE 2002: PROCEEDINGS OF THE 41ST SICE ANNUAL CONFERENCE, VOLS 1-5, 2002, : 829 - 832
  • [9] Continuous Control with a Combination of Supervised and Reinforcement Learning
    Kangin, Dmitry
    Pugeault, Nicolas
    [J]. 2018 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2018, : 163 - 170
  • [10] Reinforcement learning for continuous stochastic control problems
    Munos, R
    Bourgine, P
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 10, 1998, 10 : 1029 - 1035