Symmetric actor-critic deep reinforcement learning for cascade quadrotor flight control

被引:0
|
作者
Han, Haoran [1 ]
Cheng, Jian [1 ]
Xi, Zhilong [1 ]
Lv, Maolong [2 ]
机构
[1] Univ Elect Sci & Technol China, Sch Informat & Commun Engn, Chengdu 611731, Peoples R China
[2] Air Force Engn Univ, Air Traff Control & Nav Coll, Xian 710051, Peoples R China
关键词
Quadrotor; Flight control; Deep reinforcement learning; Symmetric actor and critic;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Even though deep reinforcement learning (DRL) has been extensively applied to quadrotor flight control to simplify parameter adjustment, it has some drawbacks in terms of control performance, such as instability and asymmetry. To address these problems, we propose an odd symmetric actor to achieve stable and symmetric control performance, and an even critic to stabilize the training process. Concretely, the bias of neural networks is eliminated, and the absolute value operation is adopted to construct the activation function. Furthermore, we devise a cascade architecture, where each module trained with DRL controls a symmetric subsystem of the quadrotor. Comparative simulations have verified the effectiveness of the proposed control scheme, which shows superiority in dealing with high -dimensional, nonlinear subsystems and disadvantage in dealing with low -dimensional, linear subsystems.
引用
收藏
页数:13
相关论文
共 50 条
  • [31] Automatic collective motion tuning using actor-critic deep reinforcement learning
    Abpeikar, Shadi
    Kasmarik, Kathryn
    Garratt, Matthew
    Hunjet, Robert
    Khan, Md Mohiuddin
    Qiu, Huanneng
    [J]. SWARM AND EVOLUTIONARY COMPUTATION, 2022, 72
  • [32] Actor-Critic reinforcement learning based on prior knowledge
    [J]. Yang, Zhenyu, 1600, Transport and Telecommunication Institute, Lomonosova street 1, Riga, LV-1019, Latvia (18):
  • [33] Estimation Error Correction in Deep Reinforcement Learning for Deterministic Actor-Critic Methods
    Saglam, Baturay
    Duran, Enes
    Cicek, Dogan C.
    Mutlu, Furkan B.
    Kozat, Suleyman S.
    [J]. 2021 IEEE 33RD INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2021), 2021, : 137 - 144
  • [34] Actor-Critic Deep Reinforcement Learning for Solving Job Shop Scheduling Problems
    Liu, Chien-Liang
    Chang, Chuan-Chin
    Tseng, Chun-Jan
    [J]. IEEE ACCESS, 2020, 8 : 71752 - 71762
  • [35] Stochastic Latent Actor-Critic: Deep Reinforcement Learning with a Latent Variable Model
    Lee, Alex X.
    Nagabandi, Anusha
    Abbeel, Pieter
    Levine, Sergey
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [36] Variational value learning in advantage actor-critic reinforcement learning
    Zhang, Yaozhong
    Han, Jiaqi
    Hu, Xiaofang
    Dan, Shihao
    [J]. 2020 CHINESE AUTOMATION CONGRESS (CAC 2020), 2020, : 1955 - 1960
  • [37] Dynamic spectrum access and sharing through actor-critic deep reinforcement learning
    Dong, Liang
    Qian, Yuchen
    Xing, Yuan
    [J]. EURASIP JOURNAL ON WIRELESS COMMUNICATIONS AND NETWORKING, 2022, 2022 (01)
  • [38] Actor-critic reinforcement learning for bidding in bilateral negotiation
    Arslan, Furkan
    Aydogan, Reyhan
    [J]. TURKISH JOURNAL OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCES, 2022, 30 (05) : 1695 - 1714
  • [39] Reinforcement learning with actor-critic for knowledge graph reasoning
    Zhang, Linli
    Li, Dewei
    Xi, Yugeng
    Jia, Shuai
    [J]. SCIENCE CHINA-INFORMATION SCIENCES, 2020, 63 (06)
  • [40] A Sandpile Model for Reliable Actor-Critic Reinforcement Learning
    Peng, Yiming
    Chen, Gang
    Zhang, Mengjie
    Pang, Shaoning
    [J]. 2017 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2017, : 4014 - 4021