Hybrid Soft Actor-Critic and Incremental Dual Heuristic Programming Reinforcement Learning for Fault-Tolerant Flight Control

被引:0
|
作者
Teirlinck, C. [1 ]
van Kampen, Erik-Jan [1 ]
机构
[1] Delft Univ Technol, Control & Simulat, POB 5058, NL-2600 GB Delft, Netherlands
来源
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Recent advancements in fault-tolerant flight control have involved model-free offline and online Reinforcement Learning (RL) algorithms in order to provide robust and adaptive control to autonomous systems. Inspired by recent work on Incremental Dual Heuristic Programming (IDHP) and Soft Actor-Critic (SAC), this research proposes a hybrid SAC-IDHP framework aiming to combine adaptive online learning from IDHP with the high complexity generalization power of SAC in controlling a fully coupled system. The hybrid framework is implemented into the inner loop of a cascaded altitude controller for a high-fidelity, six-degree-of-freedom model of the Cessna Citation II PH-LAB research aircraft. Compared to SAC-only, the SAC-IDHP hybrid demonstrates an improvement in tracking performance of 0.74%, 5.46% and 0.82% in nMAE for nominal case, longitudinal and lateral failure cases respectively. Random online policy initialization is eliminated due to identity initialization of the hybrid policy, resulting in an argument for increased safety. Additionally, robustness to biased sensor noise, initial flight condition and random critic initialization is demonstrated.
引用
收藏
页数:22
相关论文
共 50 条
  • [31] A model-based hybrid soft actor-critic deep reinforcement learning algorithm for optimal ventilator settings
    Chen, Shaotao
    Qiu, Xihe
    Tan, Xiaoyu
    Fang, Zhijun
    Jin, Yaochu
    INFORMATION SCIENCES, 2022, 611 : 47 - 64
  • [32] Adaptive Assist-as-needed Control Based on Actor-Critic Reinforcement Learning
    Zhang, Yufeng
    Li, Shuai
    Nolan, Karen J.
    Zanotto, Damiano
    2019 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2019, : 4066 - 4071
  • [33] Soft Actor-Critic Reinforcement Learning-Based Optimization for Analog Circuit Sizing
    Park, Sejin
    Choi, Youngchang
    Kang, Seokhyeong
    2023 20TH INTERNATIONAL SOC DESIGN CONFERENCE, ISOCC, 2023, : 47 - 48
  • [34] Taming chimeras in coupled oscillators using soft actor-critic based reinforcement learning
    Ding, Jianpeng
    Lei, Youming
    Small, Michael
    CHAOS, 2025, 35 (01)
  • [35] Low-Precision Reinforcement Learning: Running Soft Actor-Critic in Half Precision
    Bjorck, Johan
    Chen, Xiangyu
    De Sa, Christopher
    Gomes, Carla P.
    Weinberger, Kilian Q.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [36] A deep residual reinforcement learning algorithm based on Soft Actor-Critic for autonomous navigation
    Wen, Shuhuan
    Shu, Yili
    Rad, Ahmad
    Wen, Zeteng
    Guo, Zhengzheng
    Gong, Simeng
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 259
  • [37] Actor-Critic Reinforcement Learning for Linear Longitudinal Output Control of a Road Vehicle
    Puccetti, Luca
    Rathgeber, Christian
    Hohmann, Soeren
    2019 IEEE INTELLIGENT TRANSPORTATION SYSTEMS CONFERENCE (ITSC), 2019, : 2907 - 2913
  • [38] MULTI-STEP ACTOR-CRITIC FRAMEWORK FOR REINFORCEMENT LEARNING IN CONTINUOUS CONTROL
    Huang T.
    Chen G.
    Journal of Applied and Numerical Optimization, 2023, 5 (02): : 189 - 200
  • [39] Incremental Sliding-Mode Fault-Tolerant Flight Control
    Wang, Xuerui
    van Kampen, Erik-Jan
    Chu, Qiping
    Lu, Peng
    JOURNAL OF GUIDANCE CONTROL AND DYNAMICS, 2019, 42 (02) : 244 - 259
  • [40] Real-Time Bidding with Soft Actor-Critic Reinforcement Learning in Display Advertising
    Yakovleva, Dania
    Popov, Artem
    Filchenkov, Andrey
    PROCEEDINGS OF THE 2019 25TH CONFERENCE OF OPEN INNOVATIONS ASSOCIATION (FRUCT), 2019, : 373 - 382