Optimal path planning method based on epsilon-greedy Q-learning algorithm

被引:11
|
作者
Bulut, Vahide [1 ]
机构
[1] Izmir Katip Celebi Univ, Dept Engn Sci, TR-35620 Izmir, Turkey
关键词
Path planning; Mobile robot; Improved epsilon-greedy Q-learning; Reward function; Quintic trigonometric Bezier curve; MOBILE ROBOT;
D O I
10.1007/s40430-022-03399-w
中图分类号
TH [机械、仪表工业];
学科分类号
0802 ;
摘要
Path planning in an environment with obstacles is an ongoing problem for mobile robots. Q-learning algorithm increases its importance due to its utility in interacting with the environment. However, the size of state space and computational cost are the main parts to be improved. Hence, this paper proposes an improved epsilon-greedy Q-learning (IEGQL) algorithm to enhance efficiency and productivity regarding path length and computational cost. It is important to determine an effective reward function and adjust the agent's next action to ensure exploitation and exploration. We present a new reward function to ensure the environment's knowledge in advance for a mobile robot. Additionally, novel mathematical modeling is proposed to provide the optimal selection besides ensuring a rapid convergence. Since a mobile robot has difficulty moving through the path with sharp corners, the smooth path is formed after obtaining the optimal skeleton path. Furthermore, a real-world experiment is given based on the multi-objective function. The benchmark of the proposed IEGQL algorithm with the classical EGQL and A-star algorithms is presented. The experimental results and performance analysis indicate that the IEGQL algorithm generates the optimal path based on path length, computation time, low jerk, and staying closer to the optimal skeleton path.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Optimal path planning method based on epsilon-greedy Q-learning algorithm
    Vahide Bulut
    [J]. Journal of the Brazilian Society of Mechanical Sciences and Engineering, 2022, 44
  • [2] Optimal path planning approach based on Q-learning algorithm for mobile robots
    Maoudj, Abderraouf
    Hentout, Abdelfetah
    [J]. APPLIED SOFT COMPUTING, 2020, 97
  • [3] A Path Planning Algorithm for UAV Based on Improved Q-Learning
    Yan, Chao
    Xiang, Xiaojia
    [J]. 2018 2ND INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION SCIENCES (ICRAS), 2018, : 46 - 50
  • [4] A Path Planning Algorithm for Space Manipulator Based on Q-Learning
    Li, Taiguo
    Li, Quanhong
    Li, Wenxi
    Xia, Jiagao
    Tang, Wenhua
    Wang, Weiwen
    [J]. PROCEEDINGS OF 2019 IEEE 8TH JOINT INTERNATIONAL INFORMATION TECHNOLOGY AND ARTIFICIAL INTELLIGENCE CONFERENCE (ITAIC 2019), 2019, : 1566 - 1571
  • [5] Mobile robot path planning based on Q-learning algorithm
    Li, Shaochuan
    Wang, Xuiqing
    Hu, Liwei
    Liu, Ying
    [J]. 2019 WORLD ROBOT CONFERENCE SYMPOSIUM ON ADVANCED ROBOTICS AND AUTOMATION (WRC SARA 2019), 2019, : 160 - 165
  • [6] Coverage Path Planning Optimization Based on Q-Learning Algorithm
    Piardi, Luis
    Lima, Jose
    Pereira, Ana, I
    Costa, Paulo
    [J]. INTERNATIONAL CONFERENCE ON NUMERICAL ANALYSIS AND APPLIED MATHEMATICS (ICNAAM-2018), 2019, 2116
  • [7] The Method Based on Q-Learning Path Planning in Migrating Workflow
    Xiao, Song
    Wang, Xiao-lin
    [J]. PROCEEDINGS 2013 INTERNATIONAL CONFERENCE ON MECHATRONIC SCIENCES, ELECTRIC ENGINEERING AND COMPUTER (MEC), 2013, : 2204 - 2208
  • [8] Hybrid Path Planning Algorithm of the Mobile Agent Based on Q-Learning
    Gao, Tengteng
    Li, Caihong
    Liu, Guoming
    Guo, Na
    Wang, Di
    Li, Yongdi
    [J]. AUTOMATIC CONTROL AND COMPUTER SCIENCES, 2022, 56 (02) : 130 - 142
  • [9] Indoor Emergency Path Planning Based on the Q-Learning Optimization Algorithm
    Xu, Shenghua
    Gu, Yang
    Li, Xiaoyan
    Chen, Cai
    Hu, Yingyi
    Sang, Yu
    Jiang, Wenxing
    [J]. ISPRS INTERNATIONAL JOURNAL OF GEO-INFORMATION, 2022, 11 (01)
  • [10] PATH PLANNING OF MOBILE ROBOT BASED ON THE IMPROVED Q-LEARNING ALGORITHM
    Chen, Chaorui
    Wang, Dongshu
    [J]. INTERNATIONAL JOURNAL OF INNOVATIVE COMPUTING INFORMATION AND CONTROL, 2022, 18 (03): : 687 - 702