Explainable data-driven Q-learning control for a class of discrete-time linear autonomous systems

被引:0
|
作者
Perrusquia, Adolfo [1 ]
Zou, Mengbang [1 ]
Guo, Weisi [1 ]
机构
[1] Cranfield Univ, Sch Aerosp Transport & Mfg, Bedford MK43 0AL, England
关键词
Q-learning; State-transition function; Explainable Q-learning (XQL); Control policy; REINFORCEMENT; IDENTIFICATION;
D O I
10.1016/j.ins.2024.121283
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Explaining what a reinforcement learning (RL) control agent learns play a crucial role in the safety critical control domain. Most of the approaches in the state-of-the-art focused on imitation learning methods that uncover the hidden reward function of a given control policy. However, these approaches do not uncover what the RL agent learns effectively from the agent-environment interaction. The policy learned by the RL agent depends in how good the state transition mapping is inferred from the data. When the state transition mapping is wrongly inferred implies that the RL agent is not learning properly. This can compromise the safety of the surrounding environment and the agent itself. In this paper, we aim to uncover the elements learned by data-driven RL control agents in a special class of discrete-time linear autonomous systems. Here, the approach aims to add a new explainable dimension to data-driven control approaches to increase their trust and safe deployment. We focus on the classical data-driven Q-learning algorithm and propose an explainable Q-learning (XQL) algorithm that can be further expanded to other data-driven RL control agents. Simulation experiments are conducted to observe the effectiveness of the proposed approach under different scenarios using several discrete-time models of autonomous platforms.
引用
收藏
页数:15
相关论文
共 50 条
  • [1] Data-Driven Control for Linear Discrete-Time Delay Systems
    Rueda-Escobedo, Juan G.
    Fridman, Emilia
    Schiffer, Johannes
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2022, 67 (07) : 3321 - 3336
  • [2] Data-Driven based Iterative Learning Control for A Class of Discrete-Time Descriptor Systems
    Zhang, Daqing
    Yu, Jie
    Zhu, Baoyan
    PROCEEDINGS OF THE 35TH CHINESE CONTROL CONFERENCE 2016, 2016, : 3178 - 3182
  • [3] Data-Driven $H_{∞}$ Optimal Output Feedback Control for Linear Discrete-Time Systems Based on Off-Policy Q-Learning
    Zhang, Li
    Fan, Jialu
    Xue, Wenqian
    Lopez, Victor G.
    Li, Jinna
    Chai, Tianyou
    Lewis, Frank L.
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (07) : 3553 - 3567
  • [4] Minimax Q-learning design for H∞ control of linear discrete-time systems
    Li, Xinxing
    Xi, Lele
    Zha, Wenzhong
    Peng, Zhihong
    FRONTIERS OF INFORMATION TECHNOLOGY & ELECTRONIC ENGINEERING, 2022, 23 (03) : 438 - 451
  • [5] Improved Q-Learning Method for Linear Discrete-Time Systems
    Chen, Jian
    Wang, Jinhua
    Huang, Jie
    PROCESSES, 2020, 8 (03)
  • [6] Data-Driven Control for Linear Discrete-Time Systems with Input Saturation
    Lu, Xiaoyun
    Zhou, Dongpeng
    Chen, Wu-Hua
    Lu, Xiaomei
    CIRCUITS SYSTEMS AND SIGNAL PROCESSING, 2024, 43 (10) : 6211 - 6227
  • [7] Data-driven control of a class of discrete-time linear complex dynamical networks
    Wang, Wei
    Feng, Changyang
    Quan, Wei
    2019 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2019, : 6243 - 6245
  • [8] A Data-Driven Adaptive Iterative Learning Predictive Control for a Class of Discrete-time Nonlinear Systems
    Sun Heqing
    Hou Zhongsheng
    PROCEEDINGS OF THE 29TH CHINESE CONTROL CONFERENCE, 2010, : 5871 - 5876
  • [9] Data-driven Predictive Terminal Iterative Learning Control for a Class of Discrete-time Nonlinear Systems
    Jin Shangtai
    Hou Zhongsheng
    Chi Ronghu
    2013 32ND CHINESE CONTROL CONFERENCE (CCC), 2013, : 2992 - 2996
  • [10] A new data-driven sliding mode learning control for discrete-time MIMO linear systems
    Cao L.
    Gao S.
    Zhao D.
    International Journal of Industrial and Systems Engineering, 2022, 42 (02) : 211 - 229