Gaussian process model based reinforcement learning

被引:1
|
作者
Yoo J.H. [1 ]
机构
[1] Hankyong National University, Korea, Republic of
关键词
Gaussian Process Regression (GPR); PILCO (Probabilistic Inference for Learning Control); Reinforcement learning control system; UAV (Unmanned Aerial Vehicle);
D O I
10.5302/J.ICROS.2019.18.0221
中图分类号
学科分类号
摘要
Reinforcement learning (RL) has been a promising approach in robotics and control because data-driven learning methods can reduce system reliance on human engineering knowledge. A model-based RL autonomously learns observed dynamics based on a general flexible nonparametric approach. Probabilistic Inference for Learning COntrol (PILCO) is one of the most data-efficient model-based RL frameworks. Since PILCO sets up a Bayesian estimator problem with a Gaussian process regression, it derives a fully deterministic approximate inference for policy evaluation, which makes it computationally efficient. However, PILCO requires a task-specific scenario. If an agent is given a new goal that is different than the original training goal, PILCO should relearn its model from scratch. This paper extends PILCO to tune a linear feedback controller with a quadratic cost function, where the quadratic cost function commonly used in control systems can adjust the trade-off relationship between control input consumption and convergence rate. The suggested method is not only able to maintain the analytic and deterministic approximate inference for policy evaluation, but is also able to interpret the controller design. The suggested RL framework is applied to the control of a small quadrotor unmanned aerial vehicle (UAV) with no given dynamics. The simulation results show the convergence of the learning control performance as a function of the number of RL iterations. © ICROS 2019.
引用
收藏
页码:746 / 751
页数:5
相关论文
共 50 条
  • [21] TAG: Teacher-Advice Mechanism With Gaussian Process for Reinforcement Learning
    Lin, Ke
    Li, Duantengchuan
    Li, Yanjie
    Chen, Shiyu
    Liu, Qi
    Gao, Jianqi
    Jin, Yanrui
    Gong, Liang
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (09) : 12419 - 12433
  • [22] Model Learning with Local Gaussian Process Regression
    Nguyen-Tuong, Duy
    Seeger, Matthias
    Peters, Jan
    ADVANCED ROBOTICS, 2009, 23 (15) : 2015 - 2034
  • [23] Gaussian processes in reinforcement learning
    Rasmussen, CE
    Kuss, M
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 16, 2004, 16 : 751 - 758
  • [24] Poster Abstract: Data Efficient HVAC Control using Gaussian Process-based Reinforcement Learning
    An, Zhiyu
    Ding, Xianzhong
    Du, Wan
    PROCEEDINGS OF THE 21ST ACM CONFERENCE ON EMBEDDED NETWORKED SENSOR SYSTEMS, SENSYS 2023, 2023, : 538 - 539
  • [25] An English teaching quality evaluation model based on Gaussian process machine learning
    Qi, Shi
    Liu, Lei
    Kumar, B. Santhosh
    Prathik, A.
    EXPERT SYSTEMS, 2022, 39 (06)
  • [26] Stability of Gaussian Process Learning Based Output Feedback Model Predictive Control
    Maiworm, Michael
    Limon, Daniel
    Maria Manzano, Jose
    Findeisen, Rolf
    IFAC PAPERSONLINE, 2018, 51 (20): : 455 - 461
  • [27] Gaussian Process-Based Learning Model Predictive Control With Application to USV
    Li, Fei
    Li, Huiping
    Wu, Chao
    IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2024, 71 (12) : 16388 - 16397
  • [28] Gaussian Process Based Model-free Control with Q-Learning
    Hauser, Jan
    Pachner, Daniel
    Havlena, Vladimir
    IFAC PAPERSONLINE, 2019, 52 (11): : 236 - 243
  • [29] Model-based Bayesian Reinforcement Learning in Factored Markov Decision Process
    Wu, Bo
    Feng, Yanpeng
    Zheng, Hongyan
    JOURNAL OF COMPUTERS, 2014, 9 (04) : 845 - 850
  • [30] Context-dependent meta-control for reinforcement learning using a Dirichlet process Gaussian mixture model
    Kim, Dongjae
    Lee, Sang Wan
    2018 6TH INTERNATIONAL CONFERENCE ON BRAIN-COMPUTER INTERFACE (BCI), 2018, : 112 - 114