Q-MPC: stable and efficient reinforcement learning using model predictive control

被引:0
|
作者
Oh, Tae Hoon [1 ]
机构
[1] Kyoto Univ, Dept Chem Engn, Kyoto, Japan
来源
IFAC PAPERSONLINE | 2023年 / 56卷 / 02期
关键词
Model predictive and optimization-based control; Reinforcement learning control; Process modeling and identification; Data-based control; Learning for control; Predictive control;
D O I
10.1016/j.ifacol.2023.10.1369
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
There is a growing interest in developing an efficient data-driven control method that can be implemented into digitized manufacturing processes. Model-free reinforcement learning (RL) is a machine learning method that can directly learn the optimal control policy from the process data. However, the model-free RL shows higher cost variance than the model-based method and may require an infeasible amount of data to learn the optimal control policy. Motivated by the fact that the system identification to linear model shows high data efficiency and stable performance, this paper proposes combining the linear model predictive control (MPC) with Q-learning. This combined scheme, Q-MPC, can improve the control performance more stably and safely. For the case study, linear MPC, Q-MPC, DDPG, TD3, and SAC methods are applied to the nonlinear benchmark system, mainly focusing on the learning speed and cost variance. Copyright (c) 2023 The Authors.
引用
收藏
页码:2727 / 2732
页数:6
相关论文
共 50 条
  • [31] Shield Model Predictive Path Integral: A Computationally Efficient Robust MPC Method Using Control Barrier Functions
    Yin J.
    Dawson C.
    Fan C.
    Tsiotras P.
    IEEE Robotics and Automation Letters, 2023, 8 (11) : 7106 - 7113
  • [32] Efficient and Simple Gaussian Process Supported Stochastic Model Predictive Control for Bioreactors using HILO-MPC
    Morabito, Bruno
    Pohlodek, Johannes
    Kranert, Lena
    Espinel-Rios, Sebastian
    Findeisen, Rolf
    IFAC PAPERSONLINE, 2022, 55 (07): : 922 - 927
  • [33] Learning When to Drive in Intersections by Combining Reinforcement Learning and Model Predictive Control
    Tram, Tommy
    Batkovic, Ivo
    Ali, Mohammad
    Sjoberg, Jonas
    2019 IEEE INTELLIGENT TRANSPORTATION SYSTEMS CONFERENCE (ITSC), 2019, : 3263 - 3268
  • [35] Reinforcement learning-based model predictive control for greenhouse climate control
    Mallick, Samuel
    Airaldi, Filippo
    Dabiri, Azita
    Sun, Congcong
    De Schutter, Bart
    SMART AGRICULTURAL TECHNOLOGY, 2025, 10
  • [36] Comparison of Deep Reinforcement Learning and Model Predictive Control for Adaptive Cruise Control
    Lin, Yuan
    McPhee, John
    Azad, Nasser L.
    IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2021, 6 (02): : 221 - 231
  • [37] Reinforcement learning and model predictive control for robust embedded quadrotor guidance and control
    Colin Greatwood
    Arthur G. Richards
    Autonomous Robots, 2019, 43 : 1681 - 1693
  • [38] Reinforcement learning and model predictive control for robust embedded quadrotor guidance and control
    Greatwood, Colin
    Richards, Arthur G.
    AUTONOMOUS ROBOTS, 2019, 43 (07) : 1681 - 1693
  • [39] Tuning of Model Predictive Control for Loading Motion of Automated Excavator Using Deep Reinforcement Learning
    Ishihara S.
    Ohtsuka T.
    IEEJ Transactions on Electronics, Information and Systems, 2024, 144 (06) : 552 - 559
  • [40] Creating Predictive Haptic Feedback For Obstacle Avoidance Using a Model Predictive Control (MPC) Framework
    Balachandran, Avinash
    Brown, Matthew
    Erlien, Stephen M.
    Gerdes, J. Christian
    2015 IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), 2015, : 31 - 36