Q-MPC: stable and efficient reinforcement learning using model predictive control

被引:0
|
作者
Oh, Tae Hoon [1 ]
机构
[1] Kyoto Univ, Dept Chem Engn, Kyoto, Japan
来源
IFAC PAPERSONLINE | 2023年 / 56卷 / 02期
关键词
Model predictive and optimization-based control; Reinforcement learning control; Process modeling and identification; Data-based control; Learning for control; Predictive control;
D O I
10.1016/j.ifacol.2023.10.1369
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
There is a growing interest in developing an efficient data-driven control method that can be implemented into digitized manufacturing processes. Model-free reinforcement learning (RL) is a machine learning method that can directly learn the optimal control policy from the process data. However, the model-free RL shows higher cost variance than the model-based method and may require an infeasible amount of data to learn the optimal control policy. Motivated by the fact that the system identification to linear model shows high data efficiency and stable performance, this paper proposes combining the linear model predictive control (MPC) with Q-learning. This combined scheme, Q-MPC, can improve the control performance more stably and safely. For the case study, linear MPC, Q-MPC, DDPG, TD3, and SAC methods are applied to the nonlinear benchmark system, mainly focusing on the learning speed and cost variance. Copyright (c) 2023 The Authors.
引用
收藏
页码:2727 / 2732
页数:6
相关论文
共 50 条
  • [41] Efficient reinforcement learning: Model-based acrobot control
    Boone, G
    1997 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION - PROCEEDINGS, VOLS 1-4, 1997, : 229 - 234
  • [42] Comparison of Model Predictive and Reinforcement Learning Methods for Fault Tolerant Control
    Ahmed, Ibrahim
    Khorasgani, Hamed
    Biswas, Gautam
    IFAC PAPERSONLINE, 2018, 51 (24): : 233 - 240
  • [43] Model predictive control with stage cost shaping inspired by reinforcement learning
    Beckenbach, Lukas
    Osinenko, Pavel
    Streif, Stefan
    2019 IEEE 58TH CONFERENCE ON DECISION AND CONTROL (CDC), 2019, : 7110 - 7115
  • [44] Building Energy Management With Reinforcement Learning and Model Predictive Control: A Survey
    Zhang, Huiliang
    Seal, Sayani
    Wu, Di
    Bouffard, Francois
    Boulet, Benoit
    IEEE ACCESS, 2022, 10 : 27853 - 27862
  • [45] Safe Reinforcement Learning with Chance-constrained Model Predictive Control
    Pfrommer, Samuel
    Gautam, Tanmay
    Zhou, Alec
    Sojoudi, Somayeh
    LEARNING FOR DYNAMICS AND CONTROL CONFERENCE, VOL 168, 2022, 168
  • [46] Model predictive control of DC/DC boost converter with reinforcement learning
    Marahatta, Anup
    Rajbhandari, Yaju
    Shrestha, Ashish
    Phuyal, Sudip
    Thapa, Anup
    Korba, Petr
    HELIYON, 2022, 8 (11)
  • [47] LeTac-MPC: Learning Model Predictive Control for Tactile-Reactive Grasping
    Xu, Zhengtong
    She, Yu
    IEEE TRANSACTIONS ON ROBOTICS, 2024, 40 : 4376 - 4395
  • [48] Combining Reinforcement Learning with Model Predictive Control for On-Ramp Merging
    Lubars, Joseph
    Gupta, Harsh
    Chinchali, Sandeep
    Li, Liyun
    Raja, Adnan
    Srikant, R.
    Wu, Xinzhou
    2021 IEEE INTELLIGENT TRANSPORTATION SYSTEMS CONFERENCE (ITSC), 2021, : 942 - 947
  • [49] Comparison of Reinforcement Learning and Model Predictive Control for a Nonlinear Continuous Process
    Rajpoot, Vikas
    Munusamy, Sudhakar
    Joshi, Tanuja
    Patil, Dinesh
    Pinnamaraju, Vivek
    IFAC PAPERSONLINE, 2024, 57 : 304 - 308
  • [50] Developing a treatment for neurogenic bladder dysfunction using Model Predictive Control (MPC)
    Gomma, Hesham W.
    El-Azab, Ahmed S.
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2017, 36 : 155 - 167