Improved Reinforcement Learning Using Stability Augmentation With Application to Quadrotor Attitude Control

被引:5
|
作者
Wu, Hangxing [1 ]
Ye, Hui [1 ]
Xue, Wentao [1 ]
Yang, Xiaofei [1 ]
机构
[1] Jiangsu Univ Sci & Technol, Sch Elect & Informat, Zhenjiang 21210U, Jiangsu, Peoples R China
来源
IEEE ACCESS | 2022年 / 10卷
基金
中国国家自然科学基金;
关键词
Training; Attitude control; Rotors; Torque; Reinforcement learning; Neural networks; Stability criteria; attitude control; proximal policy optimization; quadrotor; dimension-wise clipping; stability augmentation system;
D O I
10.1109/ACCESS.2022.3185424
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Reinforcement learning (RL) has been successfully applied to motion control, without requiring accurate models and selection of control parameters. In this paper, we propose a novel RL algorithm based on proximal policy optimization algorithm with dimension-wise clipping (PPO-DWC) for attitude control of quadrotor. Firstly, dimension-wise clipping technique is introduced to solve the zero-gradient problem of the PPO algorithm, which can quickly converge while maintaining good sampling efficiency, thus improving the control performance. Moreover, following the idea of stability augmentation system (SAS), a feedback controller is designed and integrated into the environment before training the PPO controller to avoid ineffective exploration and improve the system's convergence. The eventual controller consists of two parts: the first is the result of the actor neural network in the PPO algorithm, and the second is the output of the stability augmentation feedback controller. Both of them directly use an end-to-end style of control commands to map the system state. This control architecture is applied in the attitude control of the quadrotor. The simulation results show that the quadrotor can quickly and accurately track the command and has a small steady-state error after the training by the improved PPO algorithm. Meanwhile, compared with the traditional PID controller and basic PPO algorithm, the proposed PPO-DWC algorithm with stability augmentation framework has better performance in tracking accuracy and robustness.
引用
收藏
页码:67590 / 67604
页数:15
相关论文
共 50 条
  • [1] Control of a Quadrotor With Reinforcement Learning
    Hwangbo, Jemin
    Sa, Inkyu
    Siegwart, Roland
    Hutter, Marco
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2017, 2 (04): : 2096 - 2103
  • [2] Quadrotor motion control using deep reinforcement learning
    Jiang, Zifei
    Lynch, Alan F.
    JOURNAL OF UNMANNED VEHICLE SYSTEMS, 2021, 9 (04) : 234 - 251
  • [3] Optimized Fuzzy Attitude Control of Quadrotor Unmanned Aerial Vehicle Using Adaptive Reinforcement Learning Strategy
    Wen, Guoxing
    Yu, Dengxiu
    Zhao, Yanlong
    IEEE TRANSACTIONS ON AEROSPACE AND ELECTRONIC SYSTEMS, 2024, 60 (05) : 6075 - 6083
  • [4] Quadrotor Attitude Control Based On Improved ADRC
    Xiang, Li
    Ke, Lu
    Zhao Shanyong
    2022 34TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2022, : 4091 - 4095
  • [5] Autonomous Navigation and Control of a Quadrotor Using Deep Reinforcement Learning
    Mokhtar, Mohamed
    El-Badawy, Ayman
    2023 INTERNATIONAL CONFERENCE ON UNMANNED AIRCRAFT SYSTEMS, ICUAS, 2023, : 1045 - 1052
  • [6] Control of UAV quadrotor using reinforcement learning and robust controller
    Zhang, Zizuo
    Yang, Haiyang
    Fei, Yuanyuan
    Sun, Changyin
    Yu, Yao
    IET CONTROL THEORY AND APPLICATIONS, 2023, 17 (12): : 1599 - 1610
  • [7] Quadrotor Control using Reinforcement Learning under Wind Disturbance
    Lu, Songshuo
    Li, Yanjie
    Liu, Zihan
    2023 35TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2023, : 3233 - 3240
  • [8] Adaptive Trajectory Tracking Control using Reinforcement Learning for Quadrotor
    Lou, Wenjie
    Guo, Xiao
    INTERNATIONAL JOURNAL OF ADVANCED ROBOTIC SYSTEMS, 2016, 13
  • [9] Improved dynamic stability using reinforcement learning
    Tedrake, R
    Seung, HS
    CLIMBING AND WALKING ROBOTS, 2002, : 341 - 348
  • [10] Improved Position and Attitude Tracking Control for a Quadrotor UAV
    Xiong, Jing-Jing
    Guo, Nai-Huan
    Hong, Yu-Xiang
    Zheng, En-Hui
    2019 CHINESE AUTOMATION CONGRESS (CAC2019), 2019, : 4197 - 4201