Unconstrained feedback controller design using Q-learning from noisy data

被引:0
|
作者
Kumar, Pratyush [1 ]
Rawlings, James B. [1 ]
机构
[1] Univ Calif Santa Barbara, Dept Chem Engn, Santa Barbara, CA 93106 USA
关键词
Reinforcement learning; Q-learning; Least squares policy iteration; System identification; Maximum likelihood estimation; Linear quadratic regulator; MODEL-PREDICTIVE CONTROL; REINFORCEMENT; STABILITY; MPC;
D O I
10.1016/j.compchemeng.2023.108325
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
This paper develops a novel model-free Q-learning based approach to estimate linear, unconstrained feedback controllers from noisy process data. The proposed method is based on an extension of an available approach developed to estimate the linear quadratic regulator (LQR) for linear systems with full state measurements driven by Gaussian process noise of known covariance. First, we modify the approach to treat the case of an unknown noise covariance. Then, we use the modified approach to estimate a feedback controller for linear systems with both process and measurement noise and only output measurements. We also present a model-based maximum likelihood estimation (MLE) approach to determine a linear dynamic model and noise covariances from data, which is used to construct a regulator and state estimator for comparisons in simulation studies. The performances of the model-free and model-based controller estimation approaches are compared with an example heating, ventilation, and air-conditioning (HVAC) system. We show that the proposed Q-learning approach estimates a reasonably accurate feedback controller from 24 h of noisy data. The controllers estimated using both the model-free and model-based approaches provide similar closed-loop performances with 3.5 and 2.7% losses respectively, compared to a perfect controller that uses the true dynamic model and noise covariances of the HVAC system. Finally, we give future work directions for the model-free controller design approaches by discussing some remaining advantages of the model-based approaches.
引用
收藏
页数:13
相关论文
共 50 条
  • [1] Design of a fuzzy logic controller with Evolutionary Q-Learning
    Kim, Min-Soeng
    Lee, Ju-Jang
    INTELLIGENT AUTOMATION AND SOFT COMPUTING, 2006, 12 (04): : 369 - 381
  • [2] Data-Driven Optimal Controller Design for Maglev Train: Q-Learning Method
    Xin, Liang
    Jiang, Hongwei
    Wen, Tao
    Long, Zhiqiang
    2022 34TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2022, : 1289 - 1294
  • [3] Model-free Resilient Controller Design based on Incentive Feedback Stackelberg Game and Q-learning
    Shen, Jiajun
    Li, Fengjun
    Hashemi, Morteza
    Fang, Huazhen
    2024 IEEE 7TH INTERNATIONAL CONFERENCE ON INDUSTRIAL CYBER-PHYSICAL SYSTEMS, ICPS 2024, 2024,
  • [4] Fuzzy PID Controller Design Using Q-Learning Algorithm with a Manipulated Reward Function
    Aghaei, Vahid Tavakol
    Onat, Ahmet
    Eksin, Ibrahim
    Guzelkaya, Mujde
    2015 EUROPEAN CONTROL CONFERENCE (ECC), 2015, : 2502 - 2507
  • [5] Self-Learning Fuzzy Logic Controller using Q-Learning
    Kim, Min-Soeng
    Hong, Sun-Gi
    Lee, Ju-Jang
    Journal of Advanced Computational Intelligence and Intelligent Informatics, 2000, 4 (05) : 349 - 354
  • [6] Constructing a fuzzy logic controller using evolutionary Q-Learning
    Kim, MS
    Lee, JJ
    IECON 2000: 26TH ANNUAL CONFERENCE OF THE IEEE INDUSTRIAL ELECTRONICS SOCIETY, VOLS 1-4: 21ST CENTURY TECHNOLOGIES AND INDUSTRIAL OPPORTUNITIES, 2000, : 1785 - 1790
  • [7] Reactive fuzzy controller design by Q-learning for mobile robot navigation
    张文志
    吕恬生
    Journal of Harbin Institute of Technology, 2005, (03) : 319 - 324
  • [8] Reactive fuzzy controller design by Q-learning for mobile robot navigation
    Zhang, Wen-Zhi
    Lu, Tian-Sheng
    Journal of Harbin Institute of Technology (New Series), 2005, 12 (03) : 319 - 324
  • [9] Neural Q-Learning Controller for Mobile Robot
    Ganapathy, Velappa
    Yun, Soh Chin
    Joe, Halim Kusama
    2009 IEEE/ASME INTERNATIONAL CONFERENCE ON ADVANCED INTELLIGENT MECHATRONICS, VOLS 1-3, 2009, : 863 - 868
  • [10] Deep Reinforcement Learning: From Q-Learning to Deep Q-Learning
    Tan, Fuxiao
    Yan, Pengfei
    Guan, Xinping
    NEURAL INFORMATION PROCESSING (ICONIP 2017), PT IV, 2017, 10637 : 475 - 483