An optimized Q-Learning algorithm for mobile robot local path planning

被引:7
|
作者
Zhou, Qian [1 ]
Lian, Yang [2 ,3 ]
Wu, Jiayang [1 ]
Zhu, Mengyue [1 ]
Wang, Haiyong [2 ,3 ]
Cao, Jinli [4 ]
机构
[1] Nanjing Univ Posts & Telecommun, Sch Modern Posts, Nanjing 210003, Jiangsu, Peoples R China
[2] Nanjing Univ Posts & Telecommun, Sch Comp Sci, Sch Software, Nanjing 210037, Jiangsu, Peoples R China
[3] Nanjing Univ Posts & Telecommun, Sch Cyberspace Secur, Nanjing 210037, Jiangsu, Peoples R China
[4] La Trobe Univ, Dept Comp Sci & Comp Engn, Melbourne, Australia
基金
中国国家自然科学基金;
关键词
Mobile robot; Q-Learning algorithm; Local path planning; Reinforcement learning; Adaptive learning rate;
D O I
10.1016/j.knosys.2024.111400
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The Q-Learning algorithm is a reinforcement learning technique widely used in various fields such as path planning, intelligent transportation, penetration testing, among others. It primarily involves the interaction between an agent and its environment, enabling the agent to learn an optimal strategy that maximizes cumulative rewards. Most non-agent-based path planning algorithms face challenges in exploring completely unknown environments effectively, lacking efficient perception in unfamiliar settings. Additionally, many QLearning-based path planning algorithms suffer from slow convergence and susceptibility to getting stuck in local optimal solutions. To address these issues, an optimized version of the Q-Learning algorithm (Optimized Q-Learning, O-QL) is proposed and applied to local path planning of mobile robots. O-QL introduces novel Q -table initialization methods, incorporates a new action-selection policy, and a new reward function, and adapts the Root Mean Square Propagation (RMSprop) method in the learning rate adjustment. This adjustment dynamically tunes the learning rate based on gradient changes to accelerate learning and enhance path planning efficiency. Simulation experiments are carried out in three maze environments with different complexity levels, and the performance of the algorithm in local path planning is evaluated using steps, exploration reward, learning rate change and running time. The experimental results demonstrate that O-QL exhibits improvements across all four metrics compared to existing algorithms.
引用
收藏
页数:9
相关论文
共 50 条
  • [1] Mobile robot path planning based on Q-learning algorithm
    Li, Shaochuan
    Wang, Xuiqing
    Hu, Liwei
    Liu, Ying
    [J]. 2019 WORLD ROBOT CONFERENCE SYMPOSIUM ON ADVANCED ROBOTICS AND AUTOMATION (WRC SARA 2019), 2019, : 160 - 165
  • [2] Extended Q-Learning Algorithm for Path-Planning of a Mobile Robot
    Goswami , Indrani
    Das, Pradipta Kumar
    Konar, Amit
    Janarthanan, R.
    [J]. SIMULATED EVOLUTION AND LEARNING, 2010, 6457 : 379 - +
  • [3] PATH PLANNING OF MOBILE ROBOT BASED ON THE IMPROVED Q-LEARNING ALGORITHM
    Chen, Chaorui
    Wang, Dongshu
    [J]. INTERNATIONAL JOURNAL OF INNOVATIVE COMPUTING INFORMATION AND CONTROL, 2022, 18 (03): : 687 - 702
  • [4] Dynamic Path Planning of a Mobile Robot with Improved Q-Learning algorithm
    Li, Siding
    Xu, Xin
    Zuo, Lei
    [J]. 2015 IEEE INTERNATIONAL CONFERENCE ON INFORMATION AND AUTOMATION, 2015, : 409 - 414
  • [5] A Deterministic Improved Q-Learning for Path Planning of a Mobile Robot
    Konar, Amit
    Chakraborty, Indrani Goswami
    Singh, Sapam Jitu
    Jain, Lakhmi C.
    Nagar, Atulya K.
    [J]. IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2013, 43 (05): : 1141 - 1153
  • [6] CLSQL: Improved Q-Learning Algorithm Based on Continuous Local Search Policy for Mobile Robot Path Planning
    Ma, Tian
    Lyu, Jiahao
    Yang, Jiayi
    Xi, Runtao
    Li, Yuancheng
    An, Jinpeng
    Li, Chao
    [J]. SENSORS, 2022, 22 (15)
  • [7] A Modified Q-learning Multi Robot Path Planning Algorithm
    Li, Bo
    Liang, Hongbin
    [J]. BASIC & CLINICAL PHARMACOLOGY & TOXICOLOGY, 2020, 127 : 125 - 126
  • [8] Q-learning based method of adaptive path planning for mobile robot
    Li, Yibin
    Li, Caihong
    Zhang, Zijian
    [J]. 2006 IEEE INTERNATIONAL CONFERENCE ON INFORMATION ACQUISITION, VOLS 1 AND 2, CONFERENCE PROCEEDINGS, 2006, : 983 - 987
  • [9] Synergism of Firefly Algorithm and Q-Learning for Robot Arm Path Planning
    Sadhu, Arup Kumar
    Konar, Amit
    Bhattacharjee, Tanuka
    Das, Swagatam
    [J]. SWARM AND EVOLUTIONARY COMPUTATION, 2018, 43 : 50 - 68
  • [10] Hybrid Path Planning Algorithm of the Mobile Agent Based on Q-Learning
    Gao, Tengteng
    Li, Caihong
    Liu, Guoming
    Guo, Na
    Wang, Di
    Li, Yongdi
    [J]. AUTOMATIC CONTROL AND COMPUTER SCIENCES, 2022, 56 (02) : 130 - 142