Competitive reinforcement learning in continuous control tasks

被引:0
|
作者
Abramson, M [1 ]
Pachowicz, P [1 ]
Wechsler, H [1 ]
机构
[1] George Mason Univ, Fairfax, VA 22030 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper describes a novel hybrid reinforcement learning algorithm, Sarsa Learning Vector Quantization (SLVQ), that leaves the reinforcement part intact but employs a more effective representation of the policy function using,a piecewise constant function based upon "policy prototypes." The prototypes correspond to the pattern classes induced by the Voronoi tessellation generated by self-organizing methods like Learning Vector Quantization (LVQ). The determination of the optimal policy function can be now viewed as a pattern recognition problem in the sense that the assignment of an action to a point in the phase space is similar to the assignment of a pattern class to a point in phase space. The distributed LVQ representation of the policy function automatically generates a piecewise constant tessellation of the state space and yields in a major simplification of the learning task relative to the standard reinforcement learning algorithms for whom a discontinuous table look function, has to be learned. The feasibility and comparative advantages of the new algorithm is shown on the cart centering and mountain car problems, two control problems of increased difficulty.
引用
收藏
页码:1909 / 1914
页数:6
相关论文
共 50 条
  • [31] Continuous-time on-policy neural reinforcement learning of working memory tasks
    Zambrano, Davide
    Roelfsema, Pieter R.
    Bohte, Sander M.
    [J]. 2015 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2015,
  • [32] Continuous-Time Spike-Based Reinforcement Learning for Working Memory Tasks
    Karamanis, Marios
    Zambrano, Davide
    Bohte, Sander
    [J]. ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2018, PT II, 2018, 11140 : 250 - 262
  • [33] Reinforcement Learning in Continuous Time and Space: A Stochastic Control Approach
    Wang, Haoran
    Zariphopoulou, Thaleia
    Zhou, Xun Yu
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2020, 21
  • [34] A practical Reinforcement Learning implementation approach for continuous process control
    Patel, Kalpesh M.
    [J]. COMPUTERS & CHEMICAL ENGINEERING, 2023, 174
  • [35] Reinforcement learning in continuous time and space: A stochastic control approach
    Wang, Haoran
    Zariphopoulou, Thaleia
    Zhou, Xun Yu
    [J]. Journal of Machine Learning Research, 2020, 21
  • [36] Intrinsically-motivated reinforcement learning for control with continuous actions
    de Abril, Ildefons Magrans
    Kanai, Ryota
    [J]. 2017 2ND INTERNATIONAL CONFERENCE ON INTELLIGENT INFORMATICS AND BIOMEDICAL SCIENCES (ICIIBMS), 2017, : 212 - 213
  • [37] Continuous Control of an Underground Loader Using Deep Reinforcement Learning
    Backman, Sofi
    Lindmark, Daniel
    Bodin, Kenneth
    Servin, Martin
    Mork, Joakim
    Lofgren, Hakan
    [J]. MACHINES, 2021, 9 (10)
  • [38] Continuous Control of Complex Chemical Reaction Network with Reinforcement Learning
    Alhazmi, Khalid
    Sarathy, S. Mani
    [J]. 2020 EUROPEAN CONTROL CONFERENCE (ECC 2020), 2020, : 1066 - 1068
  • [39] Reinforcement Learning with Reference Tracking Control in Continuous State Spaces
    Hall, Joseph
    Rasmussen, Carl Edward
    Maciejowski, Jan
    [J]. 2011 50TH IEEE CONFERENCE ON DECISION AND CONTROL AND EUROPEAN CONTROL CONFERENCE (CDC-ECC), 2011, : 6019 - 6024
  • [40] Continuous action reinforcement learning applied to vehicle suspension control
    Howell, MN
    Frost, GP
    Gordon, TJ
    Wu, QH
    [J]. MECHATRONICS, 1997, 7 (03) : 263 - 276