Optimal path planning method based on epsilon-greedy Q-learning algorithm

被引:11
|
作者
Bulut, Vahide [1 ]
机构
[1] Izmir Katip Celebi Univ, Dept Engn Sci, TR-35620 Izmir, Turkey
关键词
Path planning; Mobile robot; Improved epsilon-greedy Q-learning; Reward function; Quintic trigonometric Bezier curve; MOBILE ROBOT;
D O I
10.1007/s40430-022-03399-w
中图分类号
TH [机械、仪表工业];
学科分类号
0802 ;
摘要
Path planning in an environment with obstacles is an ongoing problem for mobile robots. Q-learning algorithm increases its importance due to its utility in interacting with the environment. However, the size of state space and computational cost are the main parts to be improved. Hence, this paper proposes an improved epsilon-greedy Q-learning (IEGQL) algorithm to enhance efficiency and productivity regarding path length and computational cost. It is important to determine an effective reward function and adjust the agent's next action to ensure exploitation and exploration. We present a new reward function to ensure the environment's knowledge in advance for a mobile robot. Additionally, novel mathematical modeling is proposed to provide the optimal selection besides ensuring a rapid convergence. Since a mobile robot has difficulty moving through the path with sharp corners, the smooth path is formed after obtaining the optimal skeleton path. Furthermore, a real-world experiment is given based on the multi-objective function. The benchmark of the proposed IEGQL algorithm with the classical EGQL and A-star algorithms is presented. The experimental results and performance analysis indicate that the IEGQL algorithm generates the optimal path based on path length, computation time, low jerk, and staying closer to the optimal skeleton path.
引用
收藏
页数:14
相关论文
共 50 条
  • [21] Biologically Inspired Complete Coverage Path Planning Algorithm Based on Q-Learning
    Tan, Xiangquan
    Han, Linhui
    Gong, Hao
    Wu, Qingwen
    [J]. SENSORS, 2023, 23 (10)
  • [22] Agent Maze Path Planning Based on Simulated Annealing Q-Learning Algorithm
    Mao, Zhongtian
    Wu, Zipeng
    Fang, Xiaohan
    Cheng, Songsong
    Fan, Yuan
    [J]. 2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, : 2272 - 2276
  • [23] ETQ-learning: an improved Q-learning algorithm for path planning
    Wang, Huanwei
    Jing, Jing
    Wang, Qianlv
    He, Hongqi
    Qi, Xuyan
    Lou, Rui
    [J]. INTELLIGENT SERVICE ROBOTICS, 2024, 17 (04) : 915 - 929
  • [24] A Modified Q-learning Multi Robot Path Planning Algorithm
    Li, Bo
    Liang, Hongbin
    [J]. BASIC & CLINICAL PHARMACOLOGY & TOXICOLOGY, 2020, 127 : 125 - 126
  • [25] Model based path planning using Q-Learning
    Sharma, Avinash
    Gupta, Kanika
    Kumar, Anirudha
    Sharma, Aishwarya
    Kumar, Rajesh
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON INDUSTRIAL TECHNOLOGY (ICIT), 2017, : 837 - 842
  • [26] Improved Q-Learning Algorithm Based on Flower Pollination Algorithm and Tabulation Method for Unmanned Aerial Vehicle Path Planning
    Bo, Lan
    Zhang, Tiezhu
    Zhang, Hongxin
    Yang, Jian
    Zhang, Zhen
    Zhang, Caihong
    Liu, Mingjie
    [J]. IEEE ACCESS, 2024, 12 : 104429 - 104444
  • [27] Q-learning based Path Planning Method for UAVs using Priority Shifting
    de Carvalho, Kevin B.
    de Oliveira, Iure Rosa L.
    Villa, Daniel K. D.
    Caldeira, Alexandre G.
    Sarcinelli-Filho, Mario
    Brandao, Alexandre S.
    [J]. 2022 INTERNATIONAL CONFERENCE ON UNMANNED AIRCRAFT SYSTEMS (ICUAS), 2022, : 421 - 426
  • [28] RSMDP-BASED ROBUST Q-LEARNING FOR OPTIMAL PATH PLANNING IN A DYNAMIC ENVIRONMENT
    Zhang, Yunfei
    Li, Weilin
    de Silva, Clarence W.
    [J]. INTERNATIONAL JOURNAL OF ROBOTICS & AUTOMATION, 2016, 31 (04): : 290 - 300
  • [29] UAV path planning algorithm based on Deep Q-Learning to search for a lost in the ocean
    Boulares, Mehrez
    Fehri, Afef
    Jemni, Mohamed
    [J]. ROBOTICS AND AUTONOMOUS SYSTEMS, 2024, 179
  • [30] Focused Crawler Based on Reinforcement Learning and Decaying Epsilon-Greedy Exploration Policy
    Kaleel, Parisa Begum
    Sheen, Shina
    [J]. INTERNATIONAL ARAB JOURNAL OF INFORMATION TECHNOLOGY, 2023, 20 (05) : 819 - 830