Predator-Prey Reward Based Q-Learning Coverage Path Planning for Mobile Robot

被引:4
|
作者
Zhang, Meiyan [1 ]
Cai, Wenyu [2 ]
Pang, Lingfeng [2 ]
机构
[1] Zhejiang Univ Water Resources & Elect Power, Coll Elect Engn, Hangzhou 310018, Peoples R China
[2] Hangzhou Dianzi Univ, Coll Elect & Informat, Hangzhou 310018, Peoples R China
来源
IEEE ACCESS | 2023年 / 11卷
基金
中国国家自然科学基金;
关键词
Path planning; Mobile robots; Predator prey systems; Q-learning; Planning; Partitioning algorithms; Behavioral sciences; Coverage path planning; predator-prey model; reinforcement learning; Q-learning algorithm; mobile robot; ALGORITHM; AREAS;
D O I
10.1109/ACCESS.2023.3255007
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Coverage Path Planning (CPP in short) is a basic problem for mobile robot when facing a variety of applications. Q-Learning based coverage path planning algorithms are beginning to be explored recently. To overcome the problem of traditional Q-Learning of easily falling into local optimum, in this paper, the new-type reward functions originating from Predator-Prey model are introduced into traditional Q-Learning based CPP solution, which introduces a comprehensive reward function that incorporates three rewards including Predation Avoidance Reward Function, Smoothness Reward Function and Boundary Reward Function. In addition, the influence of weighting parameters on the total reward function is discussed. Extensive simulation results and practical experiments verify that the proposed Predator-Prey reward based Q-Learning Coverage Path Planning (PP-Q-Learning based CPP in short) has better performance than traditional BCD and Q-Learning based CPP in terms of repetition ratio and turns number.
引用
收藏
页码:29673 / 29683
页数:11
相关论文
共 50 条
  • [31] Neural Q-Learning Based Mobile Robot Navigation
    Yun, Soh Chin
    Parasuraman, S.
    Ganapathy, V.
    Joe, Halim Kusuma
    [J]. MATERIALS SCIENCE AND INFORMATION TECHNOLOGY, PTS 1-8, 2012, 433-440 : 721 - +
  • [32] Synergism of Firefly Algorithm and Q-Learning for Robot Arm Path Planning
    Sadhu, Arup Kumar
    Konar, Amit
    Bhattacharjee, Tanuka
    Das, Swagatam
    [J]. SWARM AND EVOLUTIONARY COMPUTATION, 2018, 43 : 50 - 68
  • [33] A dynamic reward-enhanced Q-learning approach for efficient path planning and obstacle avoidance in mobile robotics
    Gharbi, Atef
    [J]. APPLIED COMPUTING AND INFORMATICS, 2024,
  • [34] Local Path Planning: Dynamic Window Approach With Q-Learning Considering Congestion Environments for Mobile Robot
    Kobayashi, Masato
    Zushi, Hiroka
    Nakamura, Tomoaki
    Motoi, Naoki
    [J]. IEEE ACCESS, 2023, 11 : 96733 - 96742
  • [35] Mobile robot local path planning based on Q reinforcement learning and CMAC
    Wang Zhongmin
    Yue Hong
    [J]. Proceedings of the 24th Chinese Control Conference, Vols 1 and 2, 2005, : 1494 - 1496
  • [36] PPCPP: A Predator-Prey-Based Approach to Adaptive Coverage Path Planning
    Hassan, Mahdi
    Liu, Dikai
    [J]. IEEE TRANSACTIONS ON ROBOTICS, 2020, 36 (01) : 284 - 301
  • [37] Deep recurrent Q-learning for energy-constrained coverage with a mobile robot
    Zellner, Aaron
    Dutta, Ayan
    Kulbaka, Iliya
    Sharma, Gokarna
    [J]. NEURAL COMPUTING & APPLICATIONS, 2023, 35 (26): : 19087 - 19097
  • [38] Deep recurrent Q-learning for energy-constrained coverage with a mobile robot
    Aaron Zellner
    Ayan Dutta
    Iliya Kulbaka
    Gokarna Sharma
    [J]. Neural Computing and Applications, 2023, 35 : 19087 - 19097
  • [39] Coverage Path Planning for Mobile Robot Based on Genetic Algorithm
    Wang Zhongmin
    Zhu Bo
    [J]. 2014 IEEE WORKSHOP ON ELECTRONICS, COMPUTER AND APPLICATIONS, 2014, : 732 - 735
  • [40] A novel deep learning driven robot path planning strategy: Q-learning approach
    Hu, Junli
    [J]. INTERNATIONAL JOURNAL OF COMPUTER APPLICATIONS IN TECHNOLOGY, 2023, 71 (03) : 237 - 243