Data-driven offline reinforcement learning approach for quadrotor's motion and path planning

被引:0
|
作者
ZHAO, Haoran [1 ]
FU, Hang [1 ]
YANG, Fan [1 ]
QU, Che [1 ]
ZHOU, Yaoming [1 ,2 ,3 ]
机构
[1] School of Aeronautic Science and Engineering, Beihang University, Beijing,100191, China
[2] Beijing Advanced Discipline Center for Unmanned Aircraft System, Beihang University, Beijing,100191, China
[3] Tianmushan Laboratory, Hangzhou,311115, China
来源
Chinese Journal of Aeronautics | 1600年 / 37卷 / 11期
基金
中国国家自然科学基金;
关键词
Aerial vehicle - Data driven - Data-driven learning - Markov Decision Processes - Motion and path planning - Motion-planning - Offline - Reinforcement learning approach - Reinforcement learnings - Unmanned aerial vehicle;
D O I
暂无
中图分类号
学科分类号
摘要
Non-learning based motion and path planning of an Unmanned Aerial Vehicle (UAV) is faced with low computation efficiency, mapping memory occupation and local optimization problems. This article investigates the challenge of quadrotor control using offline reinforcement learning. By establishing a data-driven learning paradigm that operates without real-environment interaction, the proposed workflow offers a safer approach than traditional reinforcement learning, making it particularly suited for UAV control in industrial scenarios. The introduced algorithm evaluates dataset uncertainty and employs a pessimistic estimation to foster offline deep reinforcement learning. Experiments highlight the algorithm's superiority over traditional online reinforcement learning methods, especially when learning from offline datasets. Furthermore, the article emphasizes the importance of a more general behavior policy. In evaluations, the trained policy demonstrated versatility by adeptly navigating diverse obstacles, underscoring its real-world applicability. © 2024
引用
收藏
页码:386 / 397
相关论文
共 50 条
  • [1] Data-driven offline reinforcement learning approach for quadrotor's motion and path planning
    Zhao, Haoran
    Fu, Hang
    Yang, Fan
    Qu, Che
    Zhou, Yaoming
    CHINESE JOURNAL OF AERONAUTICS, 2024, 37 (11) : 386 - 397
  • [2] Data-Driven Reinforcement Learning Control for Quadrotor Systems
    Dang, Ngoc Trung
    Dao, Phuong Nam
    INTERNATIONAL JOURNAL OF MECHANICAL ENGINEERING AND ROBOTICS RESEARCH, 2024, 13 (05): : 495 - 501
  • [3] Data-driven Offline Reinforcement Learning for HVAC-systems
    Blad, Christian
    Bogh, Simon
    Kallesoe, Carsten Skovmose
    ENERGY, 2022, 261
  • [4] Path Planning for Construction Machines by Offline Reinforcement Learning
    Nakayama T.
    Kashi H.
    Uchimura Y.
    IEEJ Transactions on Industry Applications, 2024, 144 (05) : 367 - 373
  • [5] Data-driven energy management for electric vehicles using offline reinforcement learning
    Wang, Yong
    Wu, Jingda
    He, Hongwen
    Wei, Zhongbao
    Sun, Fengchun
    NATURE COMMUNICATIONS, 2025, 16 (01)
  • [6] A Deep Reinforcement Learning Approach for Path Following on a Quadrotor
    Rubi, Bartomeu
    Morcego, Bernardo
    Perez, Ramon
    2020 EUROPEAN CONTROL CONFERENCE (ECC 2020), 2020, : 1092 - 1098
  • [7] Towards a Data-driven Approach to Human Preferences in Motion Planning
    Menon, Arjun
    Kacker, Pooja
    Chitta, Sachin
    2015 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2015, : 920 - 927
  • [8] Data-driven constrained reinforcement learning algorithm for path tracking control of hovercraft
    Wang, Yuanhui
    Zhou, Hua
    OCEAN ENGINEERING, 2024, 307
  • [9] Data-Driven Motion Planning: A Survey on Deep Neural Networks, Reinforcement Learning, and Large Language Model Approaches
    de Carvalho, Gabriel Peixoto
    Sawanobori, Tetsuya
    Horii, Takato
    IEEE ACCESS, 2025, 13 : 52195 - 52245
  • [10] Data-driven Learning and Modeling of AUV Operational Characteristics for Optimal Path Planning
    Edwards, Joseph R.
    Smith, Joshua
    Girard, Andrew
    Wickman, Diana
    Lermusiaux, Pierre F. J.
    Subramani, Deepak N.
    Haley, Patrick J., Jr.
    Mirabito, Chris
    Kulkarni, Chinmay S.
    Jana, Sudip
    OCEANS 2017 - ABERDEEN, 2017,