Parallel Reinforcement Learning Simulation for Visual Quadrotor Navigation

被引:2
|
作者
Saunders, Jack [1 ]
Saeedi, Sajad [2 ]
Li, Wenbin [1 ]
机构
[1] Univ Bath, Dept Comp Sci, Bath, Avon, England
[2] Toronto Metropolitan Univ, Dept Mech & Ind Engn, Toronto, ON, Canada
关键词
ENVIRONMENT;
D O I
10.1109/ICRA48891.2023.10160675
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Reinforcement learning (RL) is an agent-based approach for teaching robots to navigate within the physical world. Gathering data for RL is known to be a laborious task, and real-world experiments can be risky. Simulators facilitate the collection of training data in a quicker and more cost-effective manner. However, RL frequently requires a significant number of simulation steps for an agent to become skilful at simple tasks. This is a prevalent issue within the field of RL-based visual quadrotor navigation where state dimensions are typically very large and dynamic models are complex. Furthermore, rendering images and obtaining physical properties of the agent can be computationally expensive. To solve this, we present a simulation framework, built on AirSim, which provides efficient parallel training. Building on this framework, Ape-X is modified to incorporate parallel training of AirSim environments to make use of numerous networked computers. Through experiments we were able to achieve a reduction in training time from 3.9 hours to 11 minutes, for a toy problem, using the aforementioned framework and a total of 74 agents and two networked computers. Further details including a github repo and videos about our project, PRL4AirSim, can be found at https://sites.google.com/view/prl4airsim/home
引用
收藏
页码:1357 / 1363
页数:7
相关论文
共 50 条
  • [21] VISUAL NAVIGATION OF WHEELED MOBILE ROBOTS USING DEEP REINFORCEMENT LEARNING: SIMULATION TO REAL-TIME IMPLEMENTATION
    Nwaonumah, Ezebuugo
    Samanta, Biswanath
    PROCEEDINGS OF THE ASME DYNAMIC SYSTEMS AND CONTROL CONFERENCE, DSCC2020, VOL 1, 2020,
  • [22] Effective Deep Reinforcement Learning Setups for Multiple Goals on Visual Navigation
    Takeshi Horita, Luiz Ricardo
    Wolf, Denis Fernando
    Grassi Junior, Valdir
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [23] Autonomous UAV Visual Navigation Using an Improved Deep Reinforcement Learning
    Samma, Hussein
    El-Ferik, Sami
    IEEE ACCESS, 2024, 12 : 79967 - 79977
  • [24] Multigoal Visual Navigation With Collision Avoidance via Deep Reinforcement Learning
    Xiao, Wendong
    Yuan, Liang
    He, Li
    Ran, Teng
    Zhang, Jianbo
    Cui, Jianping
    IEEE Transactions on Instrumentation and Measurement, 2022, 71
  • [25] Visual Navigation for Biped Humanoid Robots Using Deep Reinforcement Learning
    Lobos-Tsunekawa, Kenzo
    Leiva, Francisco
    Ruiz-del-Solar, Javier
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2018, 3 (04): : 3247 - 3254
  • [26] Skill-Based Hierarchical Reinforcement Learning for Target Visual Navigation
    Wang, Shuo
    Wu, Zhihao
    Hu, Xiaobo
    Lin, Youfang
    Lv, Kai
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 8920 - 8932
  • [27] Multigoal Visual Navigation With Collision Avoidance via Deep Reinforcement Learning
    Xiao, Wendong
    Yuan, Liang
    He, Li
    Ran, Teng
    Zhang, Jianbo
    Cui, Jianping
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2022, 71
  • [28] Relation-wise transformer network and reinforcement learning for visual navigation
    He Y.
    Zhou K.
    Neural Computing and Applications, 2024, 36 (21) : 13205 - 13221
  • [29] Air Learning: a deep reinforcement learning gym for autonomous aerial robot visual navigation
    Srivatsan Krishnan
    Behzad Boroujerdian
    William Fu
    Aleksandra Faust
    Vijay Janapa Reddi
    Machine Learning, 2021, 110 : 2501 - 2540
  • [30] Air Learning: a deep reinforcement learning gym for autonomous aerial robot visual navigation
    Krishnan, Srivatsan
    Boroujerdian, Behzad
    Fu, William
    Faust, Aleksandra
    Reddi, Vijay Janapa
    MACHINE LEARNING, 2021, 110 (09) : 2501 - 2540