Learning to Navigate in Turbulent Flows With Aerial Robot Swarms: A Cooperative Deep Reinforcement Learning Approach

被引:0
|
作者
Patino, Diego [1 ]
Mayya, Siddharth [2 ]
Calderon, Juan [3 ,4 ]
Daniilidis, Kostas [1 ]
Saldana, David [5 ]
机构
[1] Univ Penn, GRASP Lab, Philadelphia, PA 19104 USA
[2] Amazon Robot, Cambridge, MA 02141 USA
[3] Univ St Tomas, Bogota 110231, Colombia
[4] Bethune Cookman Univ, Daytona Beach, FL 32114 USA
[5] Lehigh Univ, Autonomous & Intelligent Robot Lab AIRLab, Bethlehem, PA 18015 USA
关键词
Robots; Robot kinematics; Robot sensing systems; Wind; Navigation; Force; Drag; Swarm robotics; reinforcement learning; wind turbulence; machine learning for robot control; graph neural networks; NEURAL-NETWORKS; FIELDS;
D O I
10.1109/LRA.2023.3280806
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Aerial operation in turbulent environments is a challenging problem due to the chaotic behavior of the flow. This problem is made even more complex when a team of aerial robots is trying to achieve coordinated motion in turbulent wind conditions. In this letter, we present a novel multi-robot controller to navigate in turbulent flows, decoupling the trajectory-tracking control from the turbulence compensation via a nested control architecture. Unlike previous works, our method does not learn to compensate for the air-flow at a specific time and space. Instead, our method learns to compensate for the flow based on its effect on the team. This is made possible via a deep reinforcement learning approach, implemented via a Graph Convolutional Neural Network (GCNN)-based architecture, which enables robots to achieve better wind compensation by processing the spatial-temporal correlation of wind flows across the team. Our approach scales well to large robot teams -as each robot only uses information from its nearest neighbors-, and generalizes well to robot teams larger than seen in training. Simulated experiments demonstrate how information sharing improves turbulence compensation in a team of aerial robots and demonstrate the flexibility of our method over different team configurations.
引用
收藏
页码:4219 / 4226
页数:8
相关论文
共 50 条
  • [1] Learning How Pedestrians Navigate: A Deep Inverse Reinforcement Learning Approach
    Fahad, Muhammad
    Chen, Zhuo
    Guo, Yi
    [J]. 2018 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2018, : 819 - 826
  • [2] Learning to Navigate for Mobile Robot with Continual Reinforcement Learning
    Wang, Ning
    Zhang, Dingyuan
    Wang, Yong
    [J]. PROCEEDINGS OF THE 39TH CHINESE CONTROL CONFERENCE, 2020, : 3701 - 3706
  • [3] A Simple and Efficient Approach for Cooperative Incremental Learning in Robot Swarms
    Di Caro, Gianni A.
    Giusti, Alessandro
    Nagi, Jawad
    Gambardella, Luca M.
    [J]. 2013 16TH INTERNATIONAL CONFERENCE ON ADVANCED ROBOTICS (ICAR), 2013,
  • [4] Learning to navigate a crystallization model with Deep Reinforcement Learning
    Manee, Vidhyadhar
    Baratti, Roberto
    Romagnoli, Jose A.
    [J]. CHEMICAL ENGINEERING RESEARCH & DESIGN, 2022, 178 : 111 - 123
  • [5] Flocking Control of UAV Swarms with Deep Reinforcement Learning Approach
    Yan, Peng
    Bai, Chengchao
    Zheng, Hongxing
    Guo, Jifeng
    [J]. PROCEEDINGS OF 2020 3RD INTERNATIONAL CONFERENCE ON UNMANNED SYSTEMS (ICUS), 2020, : 592 - 599
  • [6] Deep reinforcement learning for turbulent drag reduction in channel flows
    Guastoni, Luca
    Rabault, Jean
    Schlatter, Philipp
    Azizpour, Hossein
    Vinuesa, Ricardo
    [J]. EUROPEAN PHYSICAL JOURNAL E, 2023, 46 (04):
  • [7] Deep reinforcement learning for turbulent drag reduction in channel flows
    Luca Guastoni
    Jean Rabault
    Philipp Schlatter
    Hossein Azizpour
    Ricardo Vinuesa
    [J]. The European Physical Journal E, 2023, 46
  • [8] Learning to Navigate in Human Environments via Deep Reinforcement Learning
    Gao, Xingyuan
    Sun, Shiying
    Zhao, Xiaoguang
    Tan, Min
    [J]. NEURAL INFORMATION PROCESSING (ICONIP 2019), PT I, 2019, 11953 : 418 - 429
  • [9] Air Learning: a deep reinforcement learning gym for autonomous aerial robot visual navigation
    Srivatsan Krishnan
    Behzad Boroujerdian
    William Fu
    Aleksandra Faust
    Vijay Janapa Reddi
    [J]. Machine Learning, 2021, 110 : 2501 - 2540
  • [10] Air Learning: a deep reinforcement learning gym for autonomous aerial robot visual navigation
    Krishnan, Srivatsan
    Boroujerdian, Behzad
    Fu, William
    Faust, Aleksandra
    Reddi, Vijay Janapa
    [J]. MACHINE LEARNING, 2021, 110 (09) : 2501 - 2540