Q-MPC: stable and efficient reinforcement learning using model predictive control

被引:0
|
作者
Oh, Tae Hoon [1 ]
机构
[1] Kyoto Univ, Dept Chem Engn, Kyoto, Japan
来源
IFAC PAPERSONLINE | 2023年 / 56卷 / 02期
关键词
Model predictive and optimization-based control; Reinforcement learning control; Process modeling and identification; Data-based control; Learning for control; Predictive control;
D O I
10.1016/j.ifacol.2023.10.1369
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
There is a growing interest in developing an efficient data-driven control method that can be implemented into digitized manufacturing processes. Model-free reinforcement learning (RL) is a machine learning method that can directly learn the optimal control policy from the process data. However, the model-free RL shows higher cost variance than the model-based method and may require an infeasible amount of data to learn the optimal control policy. Motivated by the fact that the system identification to linear model shows high data efficiency and stable performance, this paper proposes combining the linear model predictive control (MPC) with Q-learning. This combined scheme, Q-MPC, can improve the control performance more stably and safely. For the case study, linear MPC, Q-MPC, DDPG, TD3, and SAC methods are applied to the nonlinear benchmark system, mainly focusing on the learning speed and cost variance. Copyright (c) 2023 The Authors.
引用
收藏
页码:2727 / 2732
页数:6
相关论文
共 50 条
  • [21] Predictive Learning Model in Cognitive Radio using Reinforcement Learning
    Tubachi, Sharada
    Venkatesan, Mithra
    Kulkarni, A., V
    2017 IEEE INTERNATIONAL CONFERENCE ON POWER, CONTROL, SIGNALS AND INSTRUMENTATION ENGINEERING (ICPCSI), 2017, : 564 - 567
  • [22] Reinforcement Learning versus Model Predictive Control on greenhouse climate control
    Morcego, Bernardo
    Yin, Wenjie
    Boersma, Sjoerd
    van Henten, Eldert
    Puig, Vicenc
    Sun, Congcong
    COMPUTERS AND ELECTRONICS IN AGRICULTURE, 2023, 215
  • [23] Learning safety in model-based Reinforcement Learning using MPC and Gaussian Processes
    Airaldi, Filippo
    De Schutter, Bart
    Dabiri, Azita
    IFAC PAPERSONLINE, 2023, 56 (02): : 5759 - 5764
  • [24] Model-free Data-driven Predictive Control Using Reinforcement Learning
    Sawant, Shambhuraj
    Reinhardt, Dirk
    Kordabad, Arash Bahari
    Gros, Sebastien
    2023 62ND IEEE CONFERENCE ON DECISION AND CONTROL, CDC, 2023, : 4046 - 4052
  • [25] Safe Reinforcement Learning Using Robust MPC
    Zanon, Mario
    Gros, Sebastien
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2021, 66 (08) : 3638 - 3652
  • [26] Nonlinear Tire Model Approximation Using Machine Learning for Efficient Model Predictive Control
    Sousa, Lucas Castro
    Hultmann Ayala, Helon Vicente
    IEEE ACCESS, 2022, 10 : 107549 - 107562
  • [27] Newton's method for reinforcement learning and model predictive control
    Bertsekas, Dimitri
    RESULTS IN CONTROL AND OPTIMIZATION, 2022, 7
  • [28] Model Predictive Control of Quadruped Robot Based on Reinforcement Learning
    Zhang, Zhitong
    Chang, Xu
    Ma, Hongxu
    An, Honglei
    Lang, Lin
    APPLIED SCIENCES-BASEL, 2023, 13 (01):
  • [29] Autonomous boat driving system using sample-efficient model predictive control-based reinforcement learning approach
    Cui, Yunduan
    Osaki, Shigeki
    Matsubara, Takamitsu
    JOURNAL OF FIELD ROBOTICS, 2021, 38 (03) : 331 - 354
  • [30] DiffTune-MPC: Closed-Loop Learning for Model Predictive Control
    Tao, Ran
    Cheng, Sheng
    Wang, Xiaofeng
    Wang, Shenlong
    Hovakimyan, Naira
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (08): : 7294 - 7301