State-chain sequential feedback reinforcement learning for path planning of autonomous mobile robots

被引:17
|
作者
Ma, Xin [1 ]
Xu, Ya [1 ]
Sun, Guo-qiang [1 ]
Deng, Li-xia [1 ]
Li, Yi-bin [1 ]
机构
[1] Shandong Univ, Sch Control Sci & Engn, Jinan 250061, Peoples R China
基金
中国国家自然科学基金;
关键词
Path planning; Q-learning; Autonomous mobile robot; Reinforcement learning; GENETIC ALGORITHMS; INITIALIZATION; ENVIRONMENTS; OPTIMIZATION; EXPLORATION; NAVIGATION; KNOWLEDGE; STRATEGY;
D O I
10.1631/jzus.C1200226
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper deals with a new approach based on Q-learning for solving the problem of mobile robot path planning in complex unknown static environments. As a computational approach to learning through interaction with the environment, reinforcement learning algorithms have been widely used for intelligent robot control, especially in the field of autonomous mobile robots. However, the learning process is slow and cumbersome. For practical applications, rapid rates of convergence are required. Aiming at the problem of slow convergence and long learning time for Q-learning based mobile robot path planning, a state-chain sequential feedback Q-learning algorithm is proposed for quickly searching for the optimal path of mobile robots in complex unknown static environments. The state chain is built during the searching process. After one action is chosen and the reward is received, the Q-values of the state-action pairs on the previously built state chain are sequentially updated with one-step Q-learning. With the increasing number of Q-values updated after one action, the number of actual steps for convergence decreases and thus, the learning time decreases, where a step is a state transition. Extensive simulations validate the efficiency of the newly proposed approach for mobile robot path planning in complex environments. The results show that the new approach has a high convergence speed and that the robot can find the collision-free optimal path in complex unknown static environments with much shorter time, compared with the one-step Q-learning algorithm and the Q(lambda)-learning algorithm.
引用
收藏
页码:167 / 178
页数:12
相关论文
共 50 条
  • [1] State-chain sequential feedback reinforcement learning for path planning of autonomous mobile robots
    Xin Ma
    Ya Xu
    Guo-qiang Sun
    Li-xia Deng
    Yi-bin Li
    [J]. Journal of Zhejiang University SCIENCE C, 2013, 14 : 167 - 178
  • [2] State-chain sequential feedback reinforcement learning for path planning of autonomous mobile robots
    Xin MA
    Ya XU
    Guo-qiang SUN
    Li-xia DENG
    Yi-bin LI
    [J]. Frontiers of Information Technology & Electronic Engineering, 2013, 14 (03) : 167 - 178
  • [3] Path Following for Autonomous Mobile Robots with Deep Reinforcement Learning
    Cao, Yu
    Ni, Kan
    Kawaguchi, Takahiro
    Hashimoto, Seiji
    [J]. SENSORS, 2024, 24 (02)
  • [4] Path Planning for Autonomous Mobile Robots
    Bashir, Khalid
    Abbasi, Sohail
    Khokhar, Waqas Nawaz
    [J]. INTERNATIONAL JOURNAL OF COMPUTER SCIENCE AND NETWORK SECURITY, 2019, 19 (11): : 132 - 138
  • [5] Dynamic Path Planning for Mobile Robots with Deep Reinforcement Learning
    Yang, Laiyi
    Bi, Jing
    Yuan, Haitao
    [J]. IFAC PAPERSONLINE, 2022, 55 (11): : 19 - 24
  • [6] Path Planning for Mobile Robots Using Transfer Reinforcement Learning
    Zheng, Xinwang
    Zheng, Wenjie
    Du, Yong
    Li, Tiejun
    Yuan, Zhansheng
    [J]. INTERNATIONAL JOURNAL ON ARTIFICIAL INTELLIGENCE TOOLS, 2024,
  • [7] Route Planning for Autonomous Mobile Robots Using a Reinforcement Learning Algorithm
    Talaat, Fatma M. M.
    Ibrahim, Abdelhameed
    El-Kenawy, El-Sayed M.
    Abdelhamid, Abdelaziz M. A.
    Alhussan, Amel Ali
    Khafaga, Doaa Sami
    Salem, Dina Ahmed
    [J]. ACTUATORS, 2023, 12 (01)
  • [8] Path planning for mobile robots using an improved reinforcement learning scheme
    Fujisawa, S
    Kurozumi, R
    Yamamoto, T
    Suita, Y
    [J]. PROCEEDINGS OF THE 2002 IEEE INTERNATIONAL SYMPOSIUM ON INTELLIGENT CONTROL, 2002, : 67 - 74
  • [9] Path planning for mobile robots using an improved reinforcement learning scheme
    Kurozumi, R
    Fujisawa, S
    Yamamoto, T
    Suita, Y
    [J]. SICE 2002: PROCEEDINGS OF THE 41ST SICE ANNUAL CONFERENCE, VOLS 1-5, 2002, : 2178 - 2183
  • [10] Path Planning for Autonomous Mobile Robots: A Review
    Sanchez-Ibanez, Jose Ricardo
    Perez-del-Pulgar, Carlos J.
    Garcia-Cerezo, Alfonso
    [J]. SENSORS, 2021, 21 (23)