Hybrid Path Planning Algorithm of the Mobile Agent Based on Q-Learning

被引:1
|
作者
Gao, Tengteng [1 ]
Li, Caihong [1 ]
Liu, Guoming [1 ]
Guo, Na [1 ]
Wang, Di [1 ]
Li, Yongdi [1 ]
机构
[1] Shandong Univ Technol, Sch Comp Sci & Technol, Zibo 255049, Peoples R China
关键词
mobile agent; path planning; Q-learning; flower pollination algorithm; CFPA-QL algorithm; CMD-QL algorithm;
D O I
10.3103/S0146411622020043
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In the path planning using Q-learning of the mobile agent, the convergence speed is too slow. So, based on Q-learning, two hybrid algorithms are proposed to improve the above problem in this paper. One algorithm is combining Manhattan distance and Q-learning (CMD-QL); the other one is combining flower pollination algorithm and Q-learning (CFPA-QL). In the former algorithm, the Q table is firstly initialized with Manhattan distance to enhance the learning efficiency of the initial stage of Q-learning; secondly, the selection strategy of the epsilon-greedy action is improved to balance the exploration-exploitation relationship of the mobile agent's actions. In the latter algorithm, the flower pollination algorithm is first used to initialize the Q table, so that Q-learning can obtain the necessary prior information which can improve the overall learning efficiency; secondly, the epsilon-greedy strategy under the minimum value of the exploration factor is adopted, which makes effective use of the action with high value. Both algorithms have been tested under known, partially known, and unknown environments, respectively. The test results show that the CMD-QL and CFPA-QL algorithms proposed in this paper can converge to the optimal path faster than the single Q-learning method, besides the CFPA-QL algorithm has the better efficiency.
引用
收藏
页码:130 / 142
页数:13
相关论文
共 50 条
  • [31] Optimal path planning method based on epsilon-greedy Q-learning algorithm
    Vahide Bulut
    [J]. Journal of the Brazilian Society of Mechanical Sciences and Engineering, 2022, 44
  • [32] Path planning of UAV using guided enhancement Q-learning algorithm
    Zhou, Bin
    Guo, Yan
    Li, Ning
    Zhong, Xijian
    [J]. Hangkong Xuebao/Acta Aeronautica et Astronautica Sinica, 2021, 42 (09):
  • [33] Synergism of Firefly Algorithm and Q-Learning for Robot Arm Path Planning
    Sadhu, Arup Kumar
    Konar, Amit
    Bhattacharjee, Tanuka
    Das, Swagatam
    [J]. SWARM AND EVOLUTIONARY COMPUTATION, 2018, 43 : 50 - 68
  • [34] Predator-Prey Reward Based Q-Learning Coverage Path Planning for Mobile Robot
    Zhang, Meiyan
    Cai, Wenyu
    Pang, Lingfeng
    [J]. IEEE ACCESS, 2023, 11 : 29673 - 29683
  • [35] The Method Based on Q-Learning Path Planning in Migrating Workflow
    Xiao, Song
    Wang, Xiao-lin
    [J]. PROCEEDINGS 2013 INTERNATIONAL CONFERENCE ON MECHATRONIC SCIENCES, ELECTRIC ENGINEERING AND COMPUTER (MEC), 2013, : 2204 - 2208
  • [36] Ship Local Path Planning Based on Improved Q-Learning
    Gong, Ming-Fan
    Xu, Hai-Xiang
    Feng, Hui
    Wang, Yong
    Xue, Xue-Hua
    [J]. Chuan Bo Li Xue/Journal of Ship Mechanics, 2022, 26 (06): : 824 - 833
  • [37] Path-Planning of Mobile Agent using Q-Learning and Real-Time Communication in an Unfavourable Situation
    Banerjee, Dhrubojyoti
    Rakshit, Pratyusha
    Konar, Amit
    Janarthanan, Ramadoss
    [J]. PROCEEDINGS OF THE 2012 WORLD CONGRESS ON INFORMATION AND COMMUNICATION TECHNOLOGIES, 2012, : 89 - 94
  • [38] A path planning approach for mobile robots using short and safe Q-learning
    Du, He
    Hao, Bing
    Zhao, Jianshuo
    Zhang, Jiamin
    Wang, Qi
    Yuan, Qi
    [J]. PLOS ONE, 2022, 17 (09):
  • [39] Solving the optimal path planning of a mobile robot using improved Q-learning
    Low, Ee Soong
    Ong, Pauline
    Cheah, Kah Chun
    [J]. ROBOTICS AND AUTONOMOUS SYSTEMS, 2019, 115 : 143 - 161
  • [40] A Novel Hybrid Path Planning Method Based on Q-Learning and Neural Network for Robot Arm
    Abdi, Ali
    Adhikari, Dibash
    Park, Ju Hong
    [J]. APPLIED SCIENCES-BASEL, 2021, 11 (15):