Learning to Navigate in Turbulent Flows With Aerial Robot Swarms: A Cooperative Deep Reinforcement Learning Approach

被引:1
|
作者
Patino, Diego [1 ]
Mayya, Siddharth [2 ]
Calderon, Juan [3 ,4 ]
Daniilidis, Kostas [1 ]
Saldana, David [5 ]
机构
[1] Univ Penn, GRASP Lab, Philadelphia, PA 19104 USA
[2] Amazon Robot, Cambridge, MA 02141 USA
[3] Univ St Tomas, Bogota 110231, Colombia
[4] Bethune Cookman Univ, Daytona Beach, FL 32114 USA
[5] Lehigh Univ, Autonomous & Intelligent Robot Lab AIRLab, Bethlehem, PA 18015 USA
关键词
Robots; Robot kinematics; Robot sensing systems; Wind; Navigation; Force; Drag; Swarm robotics; reinforcement learning; wind turbulence; machine learning for robot control; graph neural networks; NEURAL-NETWORKS; FIELDS;
D O I
10.1109/LRA.2023.3280806
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Aerial operation in turbulent environments is a challenging problem due to the chaotic behavior of the flow. This problem is made even more complex when a team of aerial robots is trying to achieve coordinated motion in turbulent wind conditions. In this letter, we present a novel multi-robot controller to navigate in turbulent flows, decoupling the trajectory-tracking control from the turbulence compensation via a nested control architecture. Unlike previous works, our method does not learn to compensate for the air-flow at a specific time and space. Instead, our method learns to compensate for the flow based on its effect on the team. This is made possible via a deep reinforcement learning approach, implemented via a Graph Convolutional Neural Network (GCNN)-based architecture, which enables robots to achieve better wind compensation by processing the spatial-temporal correlation of wind flows across the team. Our approach scales well to large robot teams -as each robot only uses information from its nearest neighbors-, and generalizes well to robot teams larger than seen in training. Simulated experiments demonstrate how information sharing improves turbulence compensation in a team of aerial robots and demonstrate the flexibility of our method over different team configurations.
引用
收藏
页码:4219 / 4226
页数:8
相关论文
共 50 条
  • [1] Learning How Pedestrians Navigate: A Deep Inverse Reinforcement Learning Approach
    Fahad, Muhammad
    Chen, Zhuo
    Guo, Yi
    2018 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2018, : 819 - 826
  • [2] Learning to Navigate for Mobile Robot with Continual Reinforcement Learning
    Wang, Ning
    Zhang, Dingyuan
    Wang, Yong
    PROCEEDINGS OF THE 39TH CHINESE CONTROL CONFERENCE, 2020, : 3701 - 3706
  • [3] A Simple and Efficient Approach for Cooperative Incremental Learning in Robot Swarms
    Di Caro, Gianni A.
    Giusti, Alessandro
    Nagi, Jawad
    Gambardella, Luca M.
    2013 16TH INTERNATIONAL CONFERENCE ON ADVANCED ROBOTICS (ICAR), 2013,
  • [4] Learning to navigate a crystallization model with Deep Reinforcement Learning
    Manee, Vidhyadhar
    Baratti, Roberto
    Romagnoli, Jose A.
    CHEMICAL ENGINEERING RESEARCH & DESIGN, 2022, 178 : 111 - 123
  • [5] Flocking Control of UAV Swarms with Deep Reinforcement Learning Approach
    Yan, Peng
    Bai, Chengchao
    Zheng, Hongxing
    Guo, Jifeng
    PROCEEDINGS OF 2020 3RD INTERNATIONAL CONFERENCE ON UNMANNED SYSTEMS (ICUS), 2020, : 592 - 599
  • [6] Deep reinforcement learning for turbulent drag reduction in channel flows
    Guastoni, Luca
    Rabault, Jean
    Schlatter, Philipp
    Azizpour, Hossein
    Vinuesa, Ricardo
    EUROPEAN PHYSICAL JOURNAL E, 2023, 46 (04):
  • [7] Deep reinforcement learning for turbulent drag reduction in channel flows
    Luca Guastoni
    Jean Rabault
    Philipp Schlatter
    Hossein Azizpour
    Ricardo Vinuesa
    The European Physical Journal E, 2023, 46
  • [8] Learning to Navigate in Human Environments via Deep Reinforcement Learning
    Gao, Xingyuan
    Sun, Shiying
    Zhao, Xiaoguang
    Tan, Min
    NEURAL INFORMATION PROCESSING (ICONIP 2019), PT I, 2019, 11953 : 418 - 429
  • [9] Air Learning: a deep reinforcement learning gym for autonomous aerial robot visual navigation
    Srivatsan Krishnan
    Behzad Boroujerdian
    William Fu
    Aleksandra Faust
    Vijay Janapa Reddi
    Machine Learning, 2021, 110 : 2501 - 2540
  • [10] Air Learning: a deep reinforcement learning gym for autonomous aerial robot visual navigation
    Krishnan, Srivatsan
    Boroujerdian, Behzad
    Fu, William
    Faust, Aleksandra
    Reddi, Vijay Janapa
    MACHINE LEARNING, 2021, 110 (09) : 2501 - 2540