Gaussian process model based reinforcement learning

被引:1
|
作者
Yoo J.H. [1 ]
机构
[1] Hankyong National University, Korea, Republic of
关键词
Gaussian Process Regression (GPR); PILCO (Probabilistic Inference for Learning Control); Reinforcement learning control system; UAV (Unmanned Aerial Vehicle);
D O I
10.5302/J.ICROS.2019.18.0221
中图分类号
学科分类号
摘要
Reinforcement learning (RL) has been a promising approach in robotics and control because data-driven learning methods can reduce system reliance on human engineering knowledge. A model-based RL autonomously learns observed dynamics based on a general flexible nonparametric approach. Probabilistic Inference for Learning COntrol (PILCO) is one of the most data-efficient model-based RL frameworks. Since PILCO sets up a Bayesian estimator problem with a Gaussian process regression, it derives a fully deterministic approximate inference for policy evaluation, which makes it computationally efficient. However, PILCO requires a task-specific scenario. If an agent is given a new goal that is different than the original training goal, PILCO should relearn its model from scratch. This paper extends PILCO to tune a linear feedback controller with a quadratic cost function, where the quadratic cost function commonly used in control systems can adjust the trade-off relationship between control input consumption and convergence rate. The suggested method is not only able to maintain the analytic and deterministic approximate inference for policy evaluation, but is also able to interpret the controller design. The suggested RL framework is applied to the control of a small quadrotor unmanned aerial vehicle (UAV) with no given dynamics. The simulation results show the convergence of the learning control performance as a function of the number of RL iterations. © ICROS 2019.
引用
收藏
页码:746 / 751
页数:5
相关论文
共 50 条
  • [41] Model-Based Robot Control with Gaussian Process Online Learning: An Experimental Demonstration
    Tesfazgi, Samuel
    Lederer, Armin
    Kunz, Johannes F.
    Ordonez-Conejo, Alejandro J.
    Hirche, Sandra
    IFAC PAPERSONLINE, 2023, 56 (02): : 501 - 506
  • [42] Considering a learning model for communication based on reinforcement learning
    Araki, Takafumi
    Tsubone, Tadashi
    Wada, Yasuhiro
    2006 SICE-ICASE INTERNATIONAL JOINT CONFERENCE, VOLS 1-13, 2006, : 1900 - +
  • [43] A reinforcement learning-based transformed inverse model strategy for nonlinear process control
    Dutta, Debaprasad
    Upreti, Simant R.
    COMPUTERS & CHEMICAL ENGINEERING, 2023, 178
  • [44] Integration Scheme for Economic Load Dispatching and Optimization Control in Coal-Fired Plants Based on Sparse Gaussian Process Model and Deep Reinforcement Learning
    Dai, Bangwu
    Chang, Yuqing
    Wang, Fuli
    Chu, Fei
    Song, Shengjun
    INDUSTRIAL & ENGINEERING CHEMISTRY RESEARCH, 2023, 62 (39) : 16025 - 16036
  • [45] Gaussian Based Non-linear Function Approximation for Reinforcement Learning
    Haider A.
    Hawe G.
    Wang H.
    Scotney B.
    SN Computer Science, 2021, 2 (3)
  • [46] Gaussian process model based predictive control
    Kocijan, J
    Murray-Smith, R
    Rasmussen, CE
    Girard, A
    PROCEEDINGS OF THE 2004 AMERICAN CONTROL CONFERENCE, VOLS 1-6, 2004, : 2214 - 2219
  • [47] A generalized degradation model based on Gaussian process
    Wang, Zhihua
    Wu, Qiong
    Zhang, Xiongjian
    Wen, Xinlei
    Zhang, Yongbo
    Liu, Chengrui
    Fu, Huimin
    MICROELECTRONICS RELIABILITY, 2018, 85 : 207 - 214
  • [48] Distributed Multi-agent Target Search and Tracking With Gaussian Process and Reinforcement Learning
    Jigang Kim
    Dohyun Jang
    H. Jin Kim
    International Journal of Control, Automation and Systems, 2023, 21 : 3057 - 3067
  • [49] Distributed Multi-agent Target Search and Tracking With Gaussian Process and Reinforcement Learning
    Kim, Jigang
    Jang, Dohyun
    Kim, H. Jin
    INTERNATIONAL JOURNAL OF CONTROL AUTOMATION AND SYSTEMS, 2023, 21 (09) : 3057 - 3067
  • [50] Probabilistic Safeguard for Reinforcement Learning Using Safety Index Guided Gaussian Process Models
    Zhao, Weiye
    He, Tairan
    Liu, Changliu
    LEARNING FOR DYNAMICS AND CONTROL CONFERENCE, VOL 211, 2023, 211