Intelligent Path Planning of Underwater Robot Based on Reinforcement Learning

被引:20
|
作者
Yang, Jiachen [1 ]
Ni, Jingfei [1 ]
Xi, Meng [1 ]
Wen, Jiabao [1 ]
Li, Yang [1 ]
机构
[1] Tianjin Univ, Sch Elect & Informat Engn, Tianjin 300072, Peoples R China
基金
中国国家自然科学基金;
关键词
Path planning; Heuristic algorithms; Vehicle dynamics; Autonomous underwater vehicles; Reinforcement learning; Oceans; Collision avoidance; path planning; obstacle avoidance; underwater robot; NEURAL-NETWORK; OPTIMIZATION;
D O I
10.1109/TASE.2022.3190901
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
As one of the commonly used vehicles for underwater detection, underwater robots are facing a series of problems. The real underwater environment is large-scale, complex, real-time and dynamic, and many unknown obstacles may exist in the underwater environment. Under such complex conditions and lack of prior knowledge, the existing path planning methods are difficult to plan, therefore they cannot effectively meet the actual demands. In response to these problems, a three-dimensional marine environment including multiple obstacles is established with the real ocean current data in this paper, which is consistent with the actual application scenarios. Then, we propose an N-step Priority Double DQN (NPDDQN) path planning algorithm, which potently realizes obstacle avoidance in the complex environment. In addition, this study proposes an experience screening mechanism, which screens the explored positive experience and improves its reuse rate, thus efficiently improving the algorithm stability in the dynamic environment. This paper verifies the better performance of reinforcement learning compared with a variety of traditional methods in three-dimensional underwater path planning. Underwater robots based on the proposed method have good autonomy and stability, which provides a new method for path planning of underwater robots. Note to Practitioners-The goal of this study is to provide a new solution for obstacle avoidance in path planning of underwater robots, which is consistent with the dynamic and real-time demands of the real environment. Existing underwater path planning researches lack a consistent environment with the actual application, and therefore we firstly construct a three-dimensional ocean environment with real ocean current data to provide support for the algorithms. Additionally, most of the algorithms are pre-planning methods or require long-time calculation, and there is little research on obstacle avoidance. In the face of obstacle changes, underwater robots with poor adaptability will cause performance decline and even economic losses. The proposed algorithm learns through interaction with the environment, and therefore it does not require any prior experience, and has good adaptability as well as fast inference speed. Especially, in the dynamic environment, algorithm performance is difficult to guarantee due to less positive experience in exploration. The proposed experience screening mechanism improves the stability of the algorithm, so that the underwater robot maintains stable performance in different dynamic environments.
引用
收藏
页码:1983 / 1996
页数:14
相关论文
共 50 条
  • [1] The Intelligent Path Planning System of Agricultural Robot via Reinforcement Learning
    Yang, Jiachen
    Ni, Jingfei
    Li, Yang
    Wen, Jiabao
    Chen, Desheng
    [J]. SENSORS, 2022, 22 (12)
  • [2] Robot path planning based on deep reinforcement learning
    Long, Yinxin
    He, Huajin
    [J]. 2020 IEEE CONFERENCE ON TELECOMMUNICATIONS, OPTICS AND COMPUTER SCIENCE (TOCS), 2020, : 151 - 154
  • [3] Robot path planning algorithm based on reinforcement learning
    Zhang, Fuhai
    Li, Ning
    Yuan, Rupeng
    Fu, Yili
    [J]. Huazhong Keji Daxue Xuebao (Ziran Kexue Ban)/Journal of Huazhong University of Science and Technology (Natural Science Edition), 2018, 46 (12): : 65 - 70
  • [4] Robot path planning in dynamic environment based on reinforcement learning
    Zhuang, Xiao-Dong
    Meng, Qing-Chun
    Wei, Tian-Bin
    Wang, Xu-Zhu
    Tan, Rui
    Li, Xiao-Jing
    [J]. Journal of Harbin Institute of Technology (New Series), 2001, 8 (03) : 253 - 255
  • [5] Research on path planning of robot based on deep reinforcement learning
    Liu, Feng
    Chen, Chang
    Li, Zhihua
    Guan, Zhi-Hong
    Wang, Hua O.
    [J]. PROCEEDINGS OF THE 39TH CHINESE CONTROL CONFERENCE, 2020, : 3730 - 3734
  • [6] Robot path planning in dynamic environment based on reinforcement learning
    庄晓东
    孟庆春
    魏天滨
    王旭柱
    谭锐
    李筱菁
    [J]. Journal of Harbin Institute of Technology(New series), 2001, (03) : 253 - 255
  • [7] Path Planning of Cleaning Robot with Reinforcement Learning
    Moon, Woohyeon
    Park, Bumgeun
    Nengroo, Sarvar Hussain
    Kim, Taeyoung
    Har, Dongsoo
    [J]. 2022 IEEE INTERNATIONAL SYMPOSIUM ON ROBOTIC AND SENSORS ENVIRONMENTS (ROSE), 2022,
  • [8] Robot Intelligent Trajectory Planning Based on PCM Guided Reinforcement Learning
    Teng, Xiang
    Fu, Jian
    Li, Cong
    Ju, ZhaoJie
    [J]. INTELLIGENT ROBOTICS AND APPLICATIONS, ICIRA 2019, PART VI, 2019, 11745 : 342 - 355
  • [9] Intelligent trajectory planning based on reinforcement learning with KCCA inference for robot
    Fu, Jian
    Teng, Xiang
    Cao, Ce
    Lou, Ping
    [J]. Huazhong Keji Daxue Xuebao (Ziran Kexue Ban)/Journal of Huazhong University of Science and Technology (Natural Science Edition), 2019, 47 (11): : 96 - 102
  • [10] Research on Path Planning Strategy of Rescue Robot Based on Reinforcement Learning
    Shi, Ying-Ming
    Zhang, Zhiyuan
    [J]. Journal of Computers (Taiwan), 2022, 33 (03) : 187 - 194