Simulation-based evaluation of model-free reinforcement learning algorithms for quadcopter attitude control and trajectory tracking

被引:0
|
作者
Yuste, Pablo Caffyn [1 ]
Iglesias Martinez, Jose Antonio [1 ]
Sanchis de Miguel, Maria Araceli [1 ]
机构
[1] Univ Carlos III Madrid, Dept Comp Sci & Engn, Leganes, Spain
关键词
Reinforcement learning; Continuous control; Model-free; Quadcopter; SYSTEM;
D O I
10.1016/j.neucom.2024.128362
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
General use quadcopters have been under development for over a decade but many of their potential applications are still under evaluation and have not yet been adopted in many of the areas that could benefit from their use. While the current generation of quadcopters use a mature set of control algorithms, the next steps, especially as autonomous features are developed, should involve a more complex learning capability to be able to adapt to unknown circumstances in a safe and reliable way. This paper provides baseline quadcopter control models learnt using eight general reinforcement learning (RL) algorithms in a simulated environment, with the object of establishing a reference performance, both in terms of precision and generation cost, for a simple set of trajectories. Each algorithm uses a tailored set of hyperparameters while, additionally, the influence of random seeds is also studied. While not all algorithms converge in the allocated computing budget, the more complex ones are able to provide stable and precise control models. This paper recommends the use of the TD3 algorithm as a reference for comparison with new RL algorithms. Additional guidance for future work is provided based on the weaknesses identified in the learning process, especially regarding the strong dependence of agent performance on random seeds.
引用
收藏
页数:16
相关论文
共 50 条
  • [1] Reinforcement learning based model-free optimized trajectory tracking strategy design for an AUV
    Duan, Kairong
    Fong, Simon
    Chen, C. L. Philip
    [J]. NEUROCOMPUTING, 2022, 469 : 289 - 297
  • [2] Model-free Based Reinforcement Learning Control Strategy of Aircraft Attitude Systems
    Huang, Dingcui
    Hu, Jiangping
    Peng, Zhinan
    Chen, Bo
    Hao, Mingrui
    Ghosh, Bijoy Kumar
    [J]. 2020 CHINESE AUTOMATION CONGRESS (CAC 2020), 2020, : 743 - 748
  • [3] Experimental evaluation of model-free reinforcement learning algorithms for continuous HVAC control
    Biemann, Marco
    Scheller, Fabian
    Liu, Xiufeng
    Huang, Lizhen
    [J]. APPLIED ENERGY, 2021, 298
  • [4] A Model-Free Online Learning Control for Attitude Tracking of Quadrotors
    Tan, Lining
    Jin, Guodong
    Zhou, Shuhua
    Wang, Lianfeng
    [J]. APPLIED SCIENCES-BASEL, 2024, 14 (03):
  • [5] Model-Free Trajectory Optimization for Reinforcement Learning
    Akrour, Riad
    Abdolmaleki, Abbas
    Abdulsamad, Hany
    Neumann, Gerhard
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 48, 2016, 48
  • [6] Model-Free Reinforcement Learning Algorithms: A Survey
    Calisir, Sinan
    Pehlivanoglu, Meltem Kurt
    [J]. 2019 27TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2019,
  • [7] Quadcopter Trajectory Tracking and Attitude Control Based on Euler Angle Limitation
    Pratama, Billie
    Muis, Abdul
    Subiantoro, Aries
    Djemai, Mohamed
    Ben Atitallah, Rabie
    [J]. 2018 6TH INTERNATIONAL CONFERENCE ON CONTROL ENGINEERING & INFORMATION TECHNOLOGY (CEIT), 2018,
  • [8] Comparing Model-free and Model-based Algorithms for Offline Reinforcement Learning
    Swazinna, Phillip
    Udluft, Steffen
    Hein, Daniel
    Runkler, Thomas
    [J]. IFAC PAPERSONLINE, 2022, 55 (15): : 19 - 26
  • [9] Model-Free Control for Soft Manipulators based on Reinforcement Learning
    You, Xuanke
    Zhang, Yixiao
    Chen, Xiaotong
    Liu, Xinghua
    Wang, Zhanchi
    Jiang, Hao
    Chen, Xiaoping
    [J]. 2017 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2017, : 2909 - 2915
  • [10] Model-Free Emergency Frequency Control Based on Reinforcement Learning
    Chen, Chunyu
    Cui, Mingjian
    Li, Fangxing
    Yin, Shengfei
    Wang, Xinan
    [J]. IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2021, 17 (04) : 2336 - 2346