An optimized Q-Learning algorithm for mobile robot local path planning

被引:7
|
作者
Zhou, Qian [1 ]
Lian, Yang [2 ,3 ]
Wu, Jiayang [1 ]
Zhu, Mengyue [1 ]
Wang, Haiyong [2 ,3 ]
Cao, Jinli [4 ]
机构
[1] Nanjing Univ Posts & Telecommun, Sch Modern Posts, Nanjing 210003, Jiangsu, Peoples R China
[2] Nanjing Univ Posts & Telecommun, Sch Comp Sci, Sch Software, Nanjing 210037, Jiangsu, Peoples R China
[3] Nanjing Univ Posts & Telecommun, Sch Cyberspace Secur, Nanjing 210037, Jiangsu, Peoples R China
[4] La Trobe Univ, Dept Comp Sci & Comp Engn, Melbourne, Australia
基金
中国国家自然科学基金;
关键词
Mobile robot; Q-Learning algorithm; Local path planning; Reinforcement learning; Adaptive learning rate;
D O I
10.1016/j.knosys.2024.111400
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The Q-Learning algorithm is a reinforcement learning technique widely used in various fields such as path planning, intelligent transportation, penetration testing, among others. It primarily involves the interaction between an agent and its environment, enabling the agent to learn an optimal strategy that maximizes cumulative rewards. Most non-agent-based path planning algorithms face challenges in exploring completely unknown environments effectively, lacking efficient perception in unfamiliar settings. Additionally, many QLearning-based path planning algorithms suffer from slow convergence and susceptibility to getting stuck in local optimal solutions. To address these issues, an optimized version of the Q-Learning algorithm (Optimized Q-Learning, O-QL) is proposed and applied to local path planning of mobile robots. O-QL introduces novel Q -table initialization methods, incorporates a new action-selection policy, and a new reward function, and adapts the Root Mean Square Propagation (RMSprop) method in the learning rate adjustment. This adjustment dynamically tunes the learning rate based on gradient changes to accelerate learning and enhance path planning efficiency. Simulation experiments are carried out in three maze environments with different complexity levels, and the performance of the algorithm in local path planning is evaluated using steps, exploration reward, learning rate change and running time. The experimental results demonstrate that O-QL exhibits improvements across all four metrics compared to existing algorithms.
引用
收藏
页数:9
相关论文
共 50 条
  • [21] Neural Q-learning in Motion Planning for Mobile Robot
    Qin, Zheng
    Gu, Jason
    [J]. 2009 IEEE INTERNATIONAL CONFERENCE ON AUTOMATION AND LOGISTICS ( ICAL 2009), VOLS 1-3, 2009, : 1024 - 1028
  • [22] Application of artificial neural network based on Q-learning for mobile robot path planning
    Li, Caihong
    Zhang, Jingyuan
    Li, Yibin
    [J]. 2006 IEEE INTERNATIONAL CONFERENCE ON INFORMATION ACQUISITION, VOLS 1 AND 2, CONFERENCE PROCEEDINGS, 2006, : 978 - 982
  • [23] Modified Q-learning with distance metric and virtual target on path planning of mobile robot
    Low, Ee Soong
    Ong, Pauline
    Low, Cheng Yee
    Omar, Rosli
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2022, 199
  • [24] The Experience-Memory Q-Learning Algorithm for Robot Path Planning in Unknown Environment
    Zhao, Meng
    Lu, Hui
    Yang, Siyi
    Guo, Fengjuan
    [J]. IEEE ACCESS, 2020, 8 : 47824 - 47844
  • [25] A modified Q-learning algorithm for robot path planning in a digital twin assembly system
    Guo, Xiaowei
    Peng, Gongzhuang
    Meng, Yingying
    [J]. INTERNATIONAL JOURNAL OF ADVANCED MANUFACTURING TECHNOLOGY, 2022, 119 (5-6): : 3951 - 3961
  • [26] A modified Q-learning algorithm for robot path planning in a digital twin assembly system
    Xiaowei Guo
    Gongzhuang Peng
    Yingying Meng
    [J]. The International Journal of Advanced Manufacturing Technology, 2022, 119 : 3951 - 3961
  • [27] Mobile robot local path planning based on Q reinforcement learning and CMAC
    Wang Zhongmin
    Yue Hong
    [J]. Proceedings of the 24th Chinese Control Conference, Vols 1 and 2, 2005, : 1494 - 1496
  • [28] Mobile Robot Path Planning using Q-Learning with Guided Distance and Moving Target Concept
    Low, Ee Soong
    Ong, Pauline
    Low, Cheng Yee
    [J]. INTERNATIONAL JOURNAL OF INTEGRATED ENGINEERING, 2021, 13 (02): : 177 - 188
  • [29] Predator-Prey Reward Based Q-Learning Coverage Path Planning for Mobile Robot
    Zhang, Meiyan
    Cai, Wenyu
    Pang, Lingfeng
    [J]. IEEE ACCESS, 2023, 11 : 29673 - 29683
  • [30] Ship Local Path Planning Based on Improved Q-Learning
    Gong, Ming-Fan
    Xu, Hai-Xiang
    Feng, Hui
    Wang, Yong
    Xue, Xue-Hua
    [J]. Chuan Bo Li Xue/Journal of Ship Mechanics, 2022, 26 (06): : 824 - 833