Multigoal Visual Navigation With Collision Avoidance via Deep Reinforcement Learning

被引:0
|
作者
Xiao, Wendong [1 ]
Yuan, Liang [1 ,2 ,3 ]
He, Li [1 ]
Ran, Teng [1 ]
Zhang, Jianbo [1 ]
Cui, Jianping [1 ]
机构
[1] Xinjiang Univ, Sch Mech Engn, Urumqi 830046, Peoples R China
[2] Beijing Univ Chem Technol, Beijing Adv Innovat Ctr Soft Matter Sci & Engn, Beijing 100029, Peoples R China
[3] Beijing Univ Chem Technol, Coll Informat Sci & Technol, Beijing 100029, Peoples R China
基金
中国国家自然科学基金;
关键词
Navigation; Visualization; Task analysis; Trajectory; Collision avoidance; Reinforcement learning; Training; deep reinforcement learning (DRL); multigoal navigation; visual sensor;
D O I
10.1109/TIM.2022.3158384
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Learning to map the images acquired by a moving agent equipped with a camera sensor to motion commands for multigoal navigation is challenging. Most existing approaches are still struggling against collision avoidance, faster convergence, and generalization. In this article, a novel actor-critic architecture is presented to learn the optimal navigation policy. We introduce single-step reward observation and collision penalty to reshape the reinforcement learning (RL) reward function. The collision perception can be obtained by the reshaped reward function and treated as measurement information from the visual observation to avoid obstacles. Besides, expert trajectories are used to generate subgoals. A subgoal reward shaping is then proposed to accelerate policy learning with the expert knowledge of subgoals. In order to generate human-aware navigation policies, an observation-action consistency (OAC) model is introduced to ensure that the agent reaches the subgoals in turn, and moves toward the target. The whole training process is performed on a self-supervised RL approach, accompanied by an expert supervision signal. This method balances the exploration and exploitation, helping the proposed model to generalize to unseen goals. The training experiments on AI2-THOR show better performance and faster convergence speed, compared with the existing approaches. For the generalization capacity to unseen goals, the proposed method achieves the state-of-the-art success rate, with at least a 30% improvement of average episode collision.
引用
收藏
页数:9
相关论文
共 50 条
  • [21] SSRL: A Safe and Smooth Reinforcement Learning Approach for Collision Avoidance in Navigation
    Zhang, Ruixian
    Yang, Jianan
    Liang, Ye
    Lu, Shengao
    Zhang, Lixian
    [J]. 2023 2ND CONFERENCE ON FULLY ACTUATED SYSTEM THEORY AND APPLICATIONS, CFASTA, 2023, : 681 - 686
  • [22] SSRL: A Safe and Smooth Reinforcement Learning Approach for Collision Avoidance in Navigation
    Zhang, Ruixian
    Yang, Jianan
    Liang, Ye
    Lu, Shengao
    Zhang, Lixian
    [J]. Proceedings of the 2nd Conference on Fully Actuated System Theory and Applications, CFASTA 2023, 2023, : 681 - 686
  • [23] A learning method for AUV collision avoidance through deep reinforcement learning
    Xu, Jian
    Huang, Fei
    Wu, Di
    Cui, Yunfei
    Yan, Zheping
    Du, Xue
    [J]. OCEAN ENGINEERING, 2022, 260
  • [24] Smooth Trajectory Collision Avoidance through Deep Reinforcement Learning
    Song, Sirui
    Saunders, Kirk
    Yue, Ye
    Liu, Jundong
    [J]. 2022 21ST IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS, ICMLA, 2022, : 914 - 919
  • [25] An Aircraft Collision Avoidance Method Based on Deep Reinforcement Learning
    Liu, Zuocheng
    Neretin, Evgeny
    Gao, Xiaoguang
    Wan, Kaifang
    [J]. 2024 9TH INTERNATIONAL CONFERENCE ON CONTROL AND ROBOTICS ENGINEERING, ICCRE 2024, 2024, : 241 - 246
  • [26] Crowd-Aware Robot Navigation for Pedestrians with Multiple Collision Avoidance Strategies via Map-based Deep Reinforcement Learning
    Yao, Shunyi
    Chen, Guangda
    Qiu, Quecheng
    Ma, Jun
    Chen, Xiaoping
    Ji, Jianmin
    [J]. 2021 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2021, : 8144 - 8150
  • [27] Formation Control with Collision Avoidance through Deep Reinforcement Learning
    Sui, Zezhi
    Pu, Zhiqiang
    Yi, Jianqiang
    Xiong, Tianyi
    [J]. 2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [28] Ship Collision Avoidance Using Constrained Deep Reinforcement Learning
    Zhang, Rui
    Wang, Xiao
    Liu, Kezhong
    Wu, Xiaolie
    Lu, Tianyou
    Chao Zhaohui
    [J]. 2018 5TH INTERNATIONAL CONFERENCE ON BEHAVIORAL, ECONOMIC, AND SOCIO-CULTURAL COMPUTING (BESC), 2018, : 115 - 120
  • [29] Drone Navigation and Avoidance of Obstacles Through Deep Reinforcement Learning
    Cetin, Ender
    Barrado, Cristina
    Munoz, Guillem
    Macias, Miguel
    Pastor, Enric
    [J]. 2019 IEEE/AIAA 38TH DIGITAL AVIONICS SYSTEMS CONFERENCE (DASC), 2019,
  • [30] Collision-avoidance under COLREGS for unmanned surface vehicles via deep reinforcement learning
    Ma, Yong
    Zhao, Yujiao
    Wang, Yulong
    Gan, Langxiong
    Zheng, Yuanzhou
    [J]. MARITIME POLICY & MANAGEMENT, 2020, 47 (05) : 665 - 686