Real-time Obstacle Avoidance for AUV Based on Reinforcement Learning and Dynamic Window Approach

被引:0
|
作者
Shen, Yue [1 ]
Xu, Han [1 ]
Wang, Dianrui [1 ]
Zhang, Yixiao [1 ]
Yan, Tianhong [2 ]
He, Bo [1 ]
机构
[1] Ocean Univ China, Sch Informat Sci & Engn, Qingdao, Peoples R China
[2] China Jiliang Univ, Sch Mech Elect Engn, Hangzhou, Peoples R China
关键词
autonomous underwater vehicle; obstacle avoidance; dynamic window approach; Q-learning;
D O I
10.1109/IEEECONF38699.2020.9389357
中图分类号
U6 [水路运输]; P75 [海洋工程];
学科分类号
0814 ; 081505 ; 0824 ; 082401 ;
摘要
As an important tool for exploring the ocean, autonomous underwater vehicle (AUV) plays an irreplaceable role in various marine activities. Due to the complexity and uncertainty of the marine environment, AUV is required to develop in a more intelligent direction. How to ensure that AUV avoids obstacles and reaches the target point smoothly is a key research issue of the AUV. The dynamic window approach (DWA) is adopted to AUV in this paper to achieve AUV's autonomous obstacle avoidance for static obstacles. The DWA is used to search for the optimal velocity command in its admissible velocity space by maximizing the objective function, however, the weights of its objective function are constant, which makes AUV lack flexibility in complex environments, and even unable to avoid obstacles. To address the above problem, reinforcement learning is introduced to optimize DWA. Q-learning, a reinforcement learning algorithm, is used to learn the weights of the DWA's objective function, which enables appropriate weights can be selected in different environments and improves the applicability of DWA in the complex environment. Compared with the original DWA, the DWA combined with Q-learning is effective and suitable for complex obstacle environments.
引用
收藏
页数:4
相关论文
共 50 条
  • [41] Dynamic Obstacle Avoidance Method for Carrier Aircraft Based on Deep Reinforcement Learning
    Xue J.
    Kong X.
    Guo Y.
    Lu A.
    Li J.
    Wan X.
    Xu M.
    Jisuanji Fuzhu Sheji Yu Tuxingxue Xuebao/Journal of Computer-Aided Design and Computer Graphics, 2021, 33 (07): : 1102 - 1112
  • [42] Control Delay in Reinforcement Learning for Real-Time Dynamic Systems: A Memoryless Approach
    Schuitema, Erik
    Busoniu, Lucian
    Babuska, Robert
    Jonker, Pieter
    IEEE/RSJ 2010 INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS 2010), 2010, : 3226 - 3231
  • [43] Mobile robot dynamic obstacle avoidance method based on improved reinforcement learning
    Xu J.
    Shao K.
    Wang J.
    Liu X.
    Zhongguo Guanxing Jishu Xuebao/Journal of Chinese Inertial Technology, 2023, 31 (01): : 92 - 99
  • [44] Reinforcement learning-based dynamic obstacle avoidance and integration of path planning
    Jaewan Choi
    Geonhee Lee
    Chibum Lee
    Intelligent Service Robotics, 2021, 14 : 663 - 677
  • [45] Real-time planning and collision avoidance control method based on deep reinforcement learning
    Xu, Xinli
    Cai, Peng
    Cao, Yunlong
    Chu, Zhenzhong
    Zhu, Wenbo
    Zhang, Weidong
    OCEAN ENGINEERING, 2023, 281
  • [46] Toward Verifiable Real-Time Obstacle Motion Prediction for Dynamic Collision Avoidance
    Kurtz, Vince
    Lin, Hai
    2019 AMERICAN CONTROL CONFERENCE (ACC), 2019, : 2633 - 2638
  • [47] Research on AUV dynamic obstacle avoidance planning based on rolling speed obstacle method
    Zhang, Fei
    Wang, Xun
    Tang, Xiang
    Kumar, Neeraj
    Hu, Ying
    Hu, Chunlei
    TRANSACTIONS OF THE INSTITUTE OF MEASUREMENT AND CONTROL, 2024, 46 (15) : 2921 - 2933
  • [48] A reinforcement learning method for dynamic obstacle avoidance in robotic mechanisms
    Maravall, D
    De Lope, J
    COMPUTATIONAL INTELLIGENT SYSTEMS FOR APPLIED RESEARCH, 2002, : 485 - 494
  • [49] Dynamic Obstacle Avoidance and Path Planning through Reinforcement Learning
    Almazrouei, Khawla
    Kamel, Ibrahim
    Rabie, Tamer
    APPLIED SCIENCES-BASEL, 2023, 13 (14):
  • [50] Internet-based real-time obstacle avoidance of a mobile robot
    Ko, JP
    Lee, JM
    JOURNAL OF MECHANICAL SCIENCE AND TECHNOLOGY, 2005, 19 (06) : 1290 - 1303