UAV Path Planning for Wireless Data Harvesting: A Deep Reinforcement Learning Approach

被引:41
|
作者
Bayerlein, Harald [1 ]
Theile, Mirco [2 ]
Caccamo, Marco [2 ]
Gesbert, David [1 ]
机构
[1] EURECOM, Commun Syst Dept, Sophia Antipolis, France
[2] Tech Univ Munich, TUM Dept Mech Engn, Munich, Germany
基金
欧洲研究理事会;
关键词
D O I
10.1109/GLOBECOM42002.2020.9322234
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Autonomous deployment of unmanned aerial vehicles (UAVs) supporting next-generation communication networks requires efficient trajectory planning methods. We propose a new end-to-end reinforcement learning (RI) approach to UAV-enabled data collection from Internet of Things (IoT) devices in an urban environment. An autonomous drone is tasked with gathering data from distributed sensor nodes subject to limited flying time and obstacle avoidance. While previous approaches, learning and non-learning based, must perform expensive recomputations or relearn a behavior when important scenario parameters such as the number of sensors, sensor positions, or maximum flying time, change, we train a double deep Q-network (DDQN) with combined experience replay to learn a UAV control policy that generalizes over changing scenario parameters. By exploiting a multi-layer map of the environment fed through convolutional network layers to the agent, we show that our proposed network architecture enables the agent to make movement decisions for a variety of scenario parameters that balance the data collection goal with flight time efficiency and safety constraints. Considerable advantages in learning efficiency from using a map centered on the UAV's position over a non-centered map are also illustrated.
引用
收藏
页数:6
相关论文
共 50 条
  • [41] Towards Real-Time Path Planning through Deep Reinforcement Learning for a UAV in Dynamic Environments
    Chao Yan
    Xiaojia Xiang
    Chang Wang
    [J]. Journal of Intelligent & Robotic Systems, 2020, 98 : 297 - 309
  • [42] Deep Reinforcement Learning Assisted UAV Path Planning Relying on Cumulative Reward Mode and Region Segmentation
    Wang, Zhipeng
    Ng, Soon Xin
    EI-Hajjar, Mohammed
    [J]. IEEE OPEN JOURNAL OF VEHICULAR TECHNOLOGY, 2024, 5 : 737 - 751
  • [43] Deep Reinforcement Learning for Trajectory Path Planning and Distributed Inference in Resource-Constrained UAV Swarms
    Dhuheir, Marwan
    Baccour, Emna
    Erbad, Aiman
    Al-Obaidi, Sinan Sabeeh
    Hamdi, Mounir
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2023, 10 (09) : 8185 - 8201
  • [44] Adaptive Informative Path Planning Using Deep Reinforcement Learning for UAV-based Active Sensing
    Rueckin, Julius
    Jin, Liren
    Popovic, Marija
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2022), 2022, : 4473 - 4479
  • [45] UAV-Based Data Collection and Wireless Power Transfer System with Deep Reinforcement Learning
    Lee, Jaewook
    Seo, Sangwon
    Ko, Haneul
    [J]. 2023 INTERNATIONAL CONFERENCE ON INFORMATION NETWORKING, ICOIN, 2023, : 400 - 403
  • [46] UAV Control for Wireless Service Provisioning in Critical Demand Areas: A Deep Reinforcement Learning Approach
    Ho, Tai Manh
    Kim-Khoa Nguyen
    Cheriet, Mohamed
    [J]. IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2021, 70 (07) : 7138 - 7152
  • [47] A Model-free Deep Reinforcement Learning Approach for Robotic Manipulators Path Planning
    Liu, Wenxing
    Niu, Hanlin
    Mahyuddin, Muhammad Nasiruddin
    Herrmann, Guido
    Carrasco, Joaquin
    [J]. 2021 21ST INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND SYSTEMS (ICCAS 2021), 2021, : 512 - 517
  • [48] Rescue path planning for urban flood: A deep reinforcement learning-based approach
    Li, Xiao-Yan
    Wang, Xia
    [J]. RISK ANALYSIS, 2024,
  • [49] Path planning for a maritime suface ship based on Deep Reinforcement Learning and weather data
    Artusi, Eva
    Chaillan, Fabien
    Napoli, Aldo
    [J]. OCEANS 2021: SAN DIEGO - PORTO, 2021,
  • [50] UAV path design with connectivity constraint based on deep reinforcement learning
    Yu, Lin
    Wu, Fahui
    Xu, Zhihai
    Xie, Zhigang
    Yang, Dingcheng
    [J]. PHYSICAL COMMUNICATION, 2022, 52