Data-driven offline reinforcement learning approach for quadrotor's motion and path planning

被引:0
|
作者
ZHAO, Haoran [1 ]
FU, Hang [1 ]
YANG, Fan [1 ]
QU, Che [1 ]
ZHOU, Yaoming [1 ,2 ,3 ]
机构
[1] School of Aeronautic Science and Engineering, Beihang University, Beijing,100191, China
[2] Beijing Advanced Discipline Center for Unmanned Aircraft System, Beihang University, Beijing,100191, China
[3] Tianmushan Laboratory, Hangzhou,311115, China
来源
Chinese Journal of Aeronautics | 1600年 / 37卷 / 11期
基金
中国国家自然科学基金;
关键词
Aerial vehicle - Data driven - Data-driven learning - Markov Decision Processes - Motion and path planning - Motion-planning - Offline - Reinforcement learning approach - Reinforcement learnings - Unmanned aerial vehicle;
D O I
暂无
中图分类号
学科分类号
摘要
Non-learning based motion and path planning of an Unmanned Aerial Vehicle (UAV) is faced with low computation efficiency, mapping memory occupation and local optimization problems. This article investigates the challenge of quadrotor control using offline reinforcement learning. By establishing a data-driven learning paradigm that operates without real-environment interaction, the proposed workflow offers a safer approach than traditional reinforcement learning, making it particularly suited for UAV control in industrial scenarios. The introduced algorithm evaluates dataset uncertainty and employs a pessimistic estimation to foster offline deep reinforcement learning. Experiments highlight the algorithm's superiority over traditional online reinforcement learning methods, especially when learning from offline datasets. Furthermore, the article emphasizes the importance of a more general behavior policy. In evaluations, the trained policy demonstrated versatility by adeptly navigating diverse obstacles, underscoring its real-world applicability. © 2024
引用
收藏
页码:386 / 397
相关论文
共 50 条
  • [31] Autonomous Quadrotor Path Planning Through Deep Reinforcement Learning With Monocular Depth Estimation
    Khojasteh, Mahdi Shahbazi
    Salimi-Badr, Armin
    IEEE OPEN JOURNAL OF VEHICULAR TECHNOLOGY, 2025, 6 : 34 - 51
  • [32] A Synergistic Approach to Data-Driven Response Planning
    O'Neill, Marty
    Poole, Michael
    Mikler, Armin R.
    DISASTER MEDICINE AND PUBLIC HEALTH PREPAREDNESS, 2021, 15 (02) : 232 - 238
  • [33] Data-driven approach for ontology learning
    Ocampo-Guzman, Isidra
    Lopez-Arevalo, Ivan
    Sosa-Sosa, Victor
    2009 6TH INTERNATIONAL CONFERENCE ON ELECTRICAL ENGINEERING, COMPUTING SCIENCE AND AUTOMATION CONTROL (CCE 2009), 2009, : 463 - 468
  • [34] Data-Driven Economic NMPC Using Reinforcement Learning
    Gros, Sebastien
    Zanon, Mario
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2020, 65 (02) : 636 - 648
  • [35] Data-driven crowd evacuation: A reinforcement learning method
    Yao, Zhenzhen
    Zhang, Guijuan
    Lu, Dianjie
    Liu, Hong
    NEUROCOMPUTING, 2019, 366 : 314 - 327
  • [36] Formulations for Data-Driven Control Design and Reinforcement Learning
    Lee, Donghwan
    Kim, Do Wan
    2022 IEEE 17TH INTERNATIONAL CONFERENCE ON CONTROL & AUTOMATION, ICCA, 2022, : 207 - 212
  • [37] Data-driven Deep Reinforcement Learning for Automated Driving
    Prabu, Avinash
    Li, Lingxi
    Chen, Yaobin
    King, Brian
    2023 IEEE 26TH INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS, ITSC, 2023, : 3790 - 3795
  • [38] Data-Driven Robust Control Using Reinforcement Learning
    Ngo, Phuong D.
    Tejedor, Miguel
    Godtliebsen, Fred
    APPLIED SCIENCES-BASEL, 2022, 12 (04):
  • [39] Data-Driven MPC for Nonlinear Systems with Reinforcement Learning
    Li, Yiran
    Wang, Qian
    Sun, Zhongqi
    Xia, Yuanqing
    2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, : 2404 - 2409
  • [40] Past Data-Driven Adaptation in Hierarchical Reinforcement Learning
    Zhang, Sijie
    Chen, Aiguo
    Wang, Tianzi
    Zhou, Xincen
    2024 16TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND COMPUTING, ICMLC 2024, 2024, : 29 - 35