Hybrid Soft Actor-Critic and Incremental Dual Heuristic Programming Reinforcement Learning for Fault-Tolerant Flight Control

被引:0
|
作者
Teirlinck, C. [1 ]
van Kampen, Erik-Jan [1 ]
机构
[1] Delft Univ Technol, Control & Simulat, POB 5058, NL-2600 GB Delft, Netherlands
来源
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Recent advancements in fault-tolerant flight control have involved model-free offline and online Reinforcement Learning (RL) algorithms in order to provide robust and adaptive control to autonomous systems. Inspired by recent work on Incremental Dual Heuristic Programming (IDHP) and Soft Actor-Critic (SAC), this research proposes a hybrid SAC-IDHP framework aiming to combine adaptive online learning from IDHP with the high complexity generalization power of SAC in controlling a fully coupled system. The hybrid framework is implemented into the inner loop of a cascaded altitude controller for a high-fidelity, six-degree-of-freedom model of the Cessna Citation II PH-LAB research aircraft. Compared to SAC-only, the SAC-IDHP hybrid demonstrates an improvement in tracking performance of 0.74%, 5.46% and 0.82% in nMAE for nominal case, longitudinal and lateral failure cases respectively. Random online policy initialization is eliminated due to identity initialization of the hybrid policy, resulting in an argument for increased safety. Additionally, robustness to biased sensor noise, initial flight condition and random critic initialization is demonstrated.
引用
收藏
页数:22
相关论文
共 50 条
  • [1] Soft Actor-Critic Deep Reinforcement Learning for Fault-Tolerant Flight Control
    Dally, Killian
    van Kampen, Erik-Jan
    arXiv, 2022,
  • [2] Adaptive fault-tolerant control for spacecraft: A dynamic Stackelberg game approach with advantage actor-critic reinforcement learning
    Meng, Yizhen
    Liu, Chun
    Liu, Yangyang
    Tan, Longyu
    AEROSPACE SCIENCE AND TECHNOLOGY, 2024, 154
  • [3] USING ACTOR-CRITIC REINFORCEMENT LEARNING FOR CONTROL AND FLIGHT FORMATION OF QUADROTORS
    Torres, Edgar
    Xu, Lei
    Sardarmehni, Tohid
    PROCEEDINGS OF ASME 2022 INTERNATIONAL MECHANICAL ENGINEERING CONGRESS AND EXPOSITION, IMECE2022, VOL 5, 2022,
  • [4] Averaged Soft Actor-Critic for Deep Reinforcement Learning
    Ding, Feng
    Ma, Guanfeng
    Chen, Zhikui
    Gao, Jing
    Li, Peng
    COMPLEXITY, 2021, 2021
  • [5] Adaptive output-feedback fault-tolerant control for space manipulator via actor-critic learning
    Yin, Yuwan
    Ning, Xin
    Xia, Dongdong
    ADVANCES IN SPACE RESEARCH, 2025, 75 (04) : 3914 - 3932
  • [6] Symmetric actor-critic deep reinforcement learning for cascade quadrotor flight control
    Han, Haoran
    Cheng, Jian
    Xi, Zhilong
    Lv, Maolong
    NEUROCOMPUTING, 2023, 559
  • [7] Symmetric actor-critic deep reinforcement learning for cascade quadrotor flight control
    Han, Haoran
    Cheng, Jian
    Xi, Zhilong
    Lv, Maolong
    NEUROCOMPUTING, 2023, 559
  • [8] Adaptive critic fault tolerant control using dual heuristic programming
    Yen, GG
    Lima, PG
    PROCEEDING OF THE 2002 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-3, 2002, : 1814 - 1819
  • [9] Prescribed Performance Fault-Tolerant Control for Uncertain Nonlinear MIMO System Using Actor-Critic Learning Structure
    Wang, Xuerao
    Wang, Qingling
    Sun, Changyin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (09) : 4479 - 4490
  • [10] Actor-Critic Reinforcement Learning for Tracking Control in Robotics
    Pane, Yudha P.
    Nageshrao, Subramanya P.
    Babuska, Robert
    2016 IEEE 55TH CONFERENCE ON DECISION AND CONTROL (CDC), 2016, : 5819 - 5826