Autonomous UAV Navigation in Dynamic Environments with Double Deep Q-Networks

被引:8
|
作者
Yang, Yupeng [1 ]
Zhang, Kai [1 ]
Liu, Dahai [2 ]
Song, Houbing [1 ]
机构
[1] Embry Riddle Aeronaut Univ, Dept Elect Engn & Comp Sci, Daytona Beach, FL 32114 USA
[2] Embry Riddle Aeronaut Univ, Coll Aviat, Daytona Beach, FL USA
关键词
UAV; autonomous navigation; deep reinforcement learning;
D O I
10.1109/dasc50938.2020.9256455
中图分类号
V [航空、航天];
学科分类号
08 ; 0825 ;
摘要
With the rapidly increasing number and complexity of unmanned aircraft systems (UAS), enabling high-density operations becomes the most important goal for UAS operations in congested airspace. However, it is difficult to capture the global environment information such as geolocation of other unmanned aerial vehicles (UAVs) and the steep terrain in real-time. As a result, avoiding dynamic obstacles rather than static ones is challenging. Previous work demonstrates the feasibility of using traditional Q-learning to solve the navigation problem in a static environment, but this method is problematic when facing a dynamic environment because it usually causes the overestimation of action values. To address this challenge, this paper presents a framework based on double deep Q-network with priority experience replay (DDQN-PER) which allows the UAVs to navigate and avoid obstacles in a dynamic environment. The model is built upon convolutional neural networks (CNNs) whose input is raw pixels of the local known environment and whose output is an action after estimating future rewards. We set up multiple experimental scenarios with static and moving obstacles for different tasks which are ranging from single-agent navigation to multi-agent navigation. Then this model is applied to other pre-defined environments, without adjustment of the architecture or learning algorithm, to validate its generalization. Experimental results demonstrate that our proposed models can allow the UAVs reach the goal successfully in new dynamic environments.
引用
收藏
页数:7
相关论文
共 50 条
  • [21] Agent Decision Processes Using Double Deep Q-Networks plus Minimax Q-Learning
    Fitch, Natalie
    Clancy, Daniel
    [J]. 2021 IEEE AEROSPACE CONFERENCE (AEROCONF 2021), 2021,
  • [22] Autonomous UAV Trail Navigation with Obstacle Avoidance Using Deep Neural Networks
    Back, Seungho
    Cho, Gangik
    Oh, Jinwoo
    Tran, Xuan-Toa
    Oh, Hyondong
    [J]. JOURNAL OF INTELLIGENT & ROBOTIC SYSTEMS, 2020, 100 (3-4) : 1195 - 1211
  • [23] Battery Control in a Smart Energy Network using Double Dueling Deep Q-Networks
    Harrold, Daniel J. B.
    Cao, Jun
    Fan, Zhong
    [J]. 2020 IEEE PES INNOVATIVE SMART GRID TECHNOLOGIES EUROPE (ISGT-EUROPE 2020): SMART GRIDS: KEY ENABLERS OF A GREEN POWER SYSTEM, 2020, : 106 - 110
  • [24] Autonomous UAV Trail Navigation with Obstacle Avoidance Using Deep Neural Networks
    Seungho Back
    Gangik Cho
    Jinwoo Oh
    Xuan-Toa Tran
    Hyondong Oh
    [J]. Journal of Intelligent & Robotic Systems, 2020, 100 : 1195 - 1211
  • [25] Crowding Game and Deep Q-Networks for Dynamic RAN Slicing in 5G Networks
    Saad, Joe
    Khawam, Kinda
    Yassin, Mohamad
    Costanzo, Salvatore
    Boulos, Karen
    [J]. PROCEEDINGS OF THE 20TH ACM INTERNATIONAL SYMPOSIUM ON MOBILITY MANAGEMENT AND WIRELESS ACCESS, MOBIWAC 2022, 2022, : 37 - 46
  • [26] Efficient Exploration through Bayesian Deep Q-Networks
    Azizzadenesheli, Kamyar
    Brunskill, Emma
    Anandkumar, Animashree
    [J]. 2018 INFORMATION THEORY AND APPLICATIONS WORKSHOP (ITA), 2018,
  • [27] Detecting Malicious Websites by using Deep Q-Networks
    Khanh Nguyen
    Park, Younghee
    [J]. 2024 SILICON VALLEY CYBERSECURITY CONFERENCE, SVCC 2024, 2024,
  • [28] Social Attentive Deep Q-Networks for Recommender Systems
    Lei, Yu
    Wang, Zhitao
    Li, Wenjie
    Pei, Hongbin
    Dai, Quanyu
    [J]. IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2022, 34 (05) : 2443 - 2457
  • [29] Historical Best Q-Networks for Deep Reinforcement Learning
    Yu, Wenwu
    Wang, Rui
    Li, Ruiying
    Gao, Jing
    Hu, Xiaohui
    [J]. 2018 IEEE 30TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI), 2018, : 6 - 11
  • [30] DIALOGUE ENVIRONMENTS ARE DIFFERENT FROM GAMES: INVESTIGATING VARIANTS OF DEEP Q-NETWORKS FOR DIALOGUE POLICY
    Wang, Yu-An
    Chen, Yun-Nung
    [J]. 2019 IEEE AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING WORKSHOP (ASRU 2019), 2019, : 1070 - 1076