Explainable data-driven Q-learning control for a class of discrete-time linear autonomous systems

被引:0
|
作者
Perrusquia, Adolfo [1 ]
Zou, Mengbang [1 ]
Guo, Weisi [1 ]
机构
[1] Cranfield Univ, Sch Aerosp Transport & Mfg, Bedford MK43 0AL, England
关键词
Q-learning; State-transition function; Explainable Q-learning (XQL); Control policy; REINFORCEMENT; IDENTIFICATION;
D O I
10.1016/j.ins.2024.121283
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Explaining what a reinforcement learning (RL) control agent learns play a crucial role in the safety critical control domain. Most of the approaches in the state-of-the-art focused on imitation learning methods that uncover the hidden reward function of a given control policy. However, these approaches do not uncover what the RL agent learns effectively from the agent-environment interaction. The policy learned by the RL agent depends in how good the state transition mapping is inferred from the data. When the state transition mapping is wrongly inferred implies that the RL agent is not learning properly. This can compromise the safety of the surrounding environment and the agent itself. In this paper, we aim to uncover the elements learned by data-driven RL control agents in a special class of discrete-time linear autonomous systems. Here, the approach aims to add a new explainable dimension to data-driven control approaches to increase their trust and safe deployment. We focus on the classical data-driven Q-learning algorithm and propose an explainable Q-learning (XQL) algorithm that can be further expanded to other data-driven RL control agents. Simulation experiments are conducted to observe the effectiveness of the proposed approach under different scenarios using several discrete-time models of autonomous platforms.
引用
收藏
页数:15
相关论文
共 50 条
  • [21] Data-Driven Superstabilizing Control of Error-in-Variables Discrete-Time Linear Systems
    Miller, Jared
    Dai, Tianyu
    Sznaier, Mario
    2022 IEEE 61ST CONFERENCE ON DECISION AND CONTROL (CDC), 2022, : 4924 - 4929
  • [22] Q-Learning for Continuous-Time Linear Systems: A Data-Driven Implementation of the Kleinman Algorithm
    Possieri, Corrado
    Sassano, Mario
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2022, 52 (10): : 6487 - 6497
  • [23] Data-driven tracking control approach for linear systems by on-policy Q-learning approach
    Zhang Yihan
    Mao Zhenfei
    Li Jinna
    16TH IEEE INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION, ROBOTICS AND VISION (ICARCV 2020), 2020, : 1066 - 1070
  • [24] Data-driven control for discrete-time piecewise affine systems✩
    Wang, Meng
    Qiu, Jianbin
    Yan, Huaicheng
    Tian, Yongxiao
    Li, Zhichen
    AUTOMATICA, 2023, 155
  • [25] Adaptive iterative learning control based on IF–THEN rules and data-driven scheme for a class of nonlinear discrete-time systems
    Chidentree Treesatayapun
    Soft Computing, 2018, 22 : 487 - 497
  • [26] Data-driven model-free sliding mode learning control for a class of discrete-time nonlinear systems
    Cao, Lei
    Gao, Shouli
    Zhao, Dongya
    TRANSACTIONS OF THE INSTITUTE OF MEASUREMENT AND CONTROL, 2020, 42 (13) : 2533 - 2547
  • [27] Data-Driven MFAC for a Class of Discrete-Time Nonlinear Systems With RBFNN
    Zhu, Yuanming
    Hou, Zhongsheng
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2014, 25 (05) : 1013 - 1020
  • [28] Data-Driven Finite-Time Control for Discrete-Time Linear Time-Invariant Systems
    Li, Jinjiang
    Liu, Tao
    Liu, Tengfei
    2023 62ND IEEE CONFERENCE ON DECISION AND CONTROL, CDC, 2023, : 1595 - 1600
  • [29] A Data-Driven ILC Framework for a Class of Nonlinear Discrete-Time Systems
    Yu, Xian
    Hou, Zhongsheng
    Polycarpou, Marios M.
    IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (07) : 6143 - 6157
  • [30] Data-driven disturbance compensation control for discrete-time systems based on reinforcement learning
    Li, Lanyue
    Li, Jinna
    Cao, Jiangtao
    INTERNATIONAL JOURNAL OF ADAPTIVE CONTROL AND SIGNAL PROCESSING, 2024,