Parallel Reinforcement Learning Simulation for Visual Quadrotor Navigation

被引:2
|
作者
Saunders, Jack [1 ]
Saeedi, Sajad [2 ]
Li, Wenbin [1 ]
机构
[1] Univ Bath, Dept Comp Sci, Bath, Avon, England
[2] Toronto Metropolitan Univ, Dept Mech & Ind Engn, Toronto, ON, Canada
关键词
ENVIRONMENT;
D O I
10.1109/ICRA48891.2023.10160675
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Reinforcement learning (RL) is an agent-based approach for teaching robots to navigate within the physical world. Gathering data for RL is known to be a laborious task, and real-world experiments can be risky. Simulators facilitate the collection of training data in a quicker and more cost-effective manner. However, RL frequently requires a significant number of simulation steps for an agent to become skilful at simple tasks. This is a prevalent issue within the field of RL-based visual quadrotor navigation where state dimensions are typically very large and dynamic models are complex. Furthermore, rendering images and obtaining physical properties of the agent can be computationally expensive. To solve this, we present a simulation framework, built on AirSim, which provides efficient parallel training. Building on this framework, Ape-X is modified to incorporate parallel training of AirSim environments to make use of numerous networked computers. Through experiments we were able to achieve a reduction in training time from 3.9 hours to 11 minutes, for a toy problem, using the aforementioned framework and a total of 74 agents and two networked computers. Further details including a github repo and videos about our project, PRL4AirSim, can be found at https://sites.google.com/view/prl4airsim/home
引用
下载
收藏
页码:1357 / 1363
页数:7
相关论文
共 50 条
  • [41] Quadrotor motion control using deep reinforcement learning
    Jiang, Zifei
    Lynch, Alan F.
    JOURNAL OF UNMANNED VEHICLE SYSTEMS, 2021, 9 (04) : 234 - 251
  • [42] Controller Design for Quadrotor UAVs using Reinforcement Learning
    Bou-Ammar, Haitham
    Voos, Holger
    Ertel, Wolfgang
    2010 IEEE INTERNATIONAL CONFERENCE ON CONTROL APPLICATIONS, 2010, : 2130 - 2135
  • [43] Supplementary Reinforcement Learning Controller Designed for Quadrotor UAVs
    Lin, Xiaobo
    Yu, Yao
    Sun, Chang-Yin
    IEEE ACCESS, 2019, 7 : 26422 - 26431
  • [44] A Reinforcement Learning Approach for Autonomous Control and Landing of a Quadrotor
    Vankadari, Madhu Babu
    Das, Kaushik
    Shinde, Chinmay
    Kumar, Swagat
    2018 INTERNATIONAL CONFERENCE ON UNMANNED AIRCRAFT SYSTEMS (ICUAS), 2018, : 676 - 683
  • [45] Inclined Quadrotor Landing using Deep Reinforcement Learning
    Kooi, Jacob E.
    Babuska, Robert
    2021 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2021, : 2361 - 2368
  • [46] Hybrid Reinforcement Learning Control for a Micro Quadrotor Flight
    Yoo, Jaehyun
    Jang, Dohyun
    Kim, H. Jin
    Johansson, Karl H.
    IEEE CONTROL SYSTEMS LETTERS, 2021, 5 (02): : 505 - 510
  • [47] A Comparison of Visual Navigation Approaches based on Localization and Reinforcement Learning in Virtual and Real Environments
    Rosano, Marco
    Furnari, Antonino
    Gulino, Luigi
    Farinella, Giovanni Maria
    PROCEEDINGS OF THE 15TH INTERNATIONAL JOINT CONFERENCE ON COMPUTER VISION, IMAGING AND COMPUTER GRAPHICS THEORY AND APPLICATIONS, VOL 5: VISAPP, 2020, : 628 - 635
  • [48] Towards Generalization in Target-Driven Visual Navigation by Using Deep Reinforcement Learning
    Devo, Alessandro
    Mezzetti, Giacomo
    Costante, Gabriele
    Fravolini, Mario L.
    Valigi, Paolo
    IEEE TRANSACTIONS ON ROBOTICS, 2020, 36 (05) : 1546 - 1561
  • [49] Optimizing Reinforcement Learning-Based Visual Navigation for Resource-Constrained Devices
    Vijetha, U.
    Geetha, V.
    IEEE ACCESS, 2023, 11 : 125648 - 125663
  • [50] Deep Reinforcement Learning Visual Navigation Model Integrating Memory- prediction Mechanism
    Xiao, Qian
    Yi, Pengfei
    Liu, Rui
    Dong, Jing
    Zhou, Dongsheng
    Zhang, Qiang
    PROCEEDINGS OF THE 2021 IEEE 24TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN (CSCWD), 2021, : 109 - 114