Flexible Route Planning for Multiple Mobile Robots by Combining Q-Learning and Graph Search Algorithm

被引:2
|
作者
Kawabe, Tomoya [1 ]
Nishi, Tatsushi [1 ]
Liu, Ziang [1 ]
机构
[1] Okayama Univ, Grad Sch Nat Sci & Technol, 3 1 1 Tsushima naka, Kita ku, Okayama 7008530, Japan
来源
APPLIED SCIENCES-BASEL | 2023年 / 13卷 / 03期
基金
日本学术振兴会;
关键词
Automated Guided Vehicles; route planning; Q-learning; reinforcement learning; transportation; NET DECOMPOSITION APPROACH; AGV; COORDINATION; DESIGN; TIME;
D O I
10.3390/app13031879
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
The use of multiple mobile robots has grown significantly over the past few years in logistics, manufacturing and public services. Conflict-free route planning is one of the major research challenges for such mobile robots. Optimization methods such as graph search algorithms are used extensively to solve route planning problems. Those methods can assure the quality of solutions, however, they are not flexible to deal with unexpected situations. In this article, we propose a flexible route planning method that combines the reinforcement learning algorithm and a graph search algorithm for conflict-free route planning problems for multiple robots. In the proposed method, Q-learning, a reinforcement algorithm, is applied to avoid collisions using off-line learning with a limited state space to reduce the total learning time. Each vehicle independently finds the shortest route using the A* algorithm, and Q-learning is used to avoid collisions. The effectiveness of the proposed method is examined by comparing it with conventional methods in terms of computation time and the quality of solutions. Computational results show that for dynamic transportation problems, the proposed method can generate the solutions with approximately 10% of the computation time compared to the conventional Q-learning approach. We found that the required computation time is linearly increased with respect to the number of vehicles and nodes in the problems.
引用
收藏
页数:21
相关论文
共 50 条
  • [1] Path planning of mobile robots with Q-learning
    Cetin, Halil
    Durdu, Akif
    2014 22ND SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2014, : 2162 - 2165
  • [2] Optimal path planning approach based on Q-learning algorithm for mobile robots
    Maoudj, Abderraouf
    Hentout, Abdelfetah
    APPLIED SOFT COMPUTING, 2020, 97
  • [3] Mobile robot path planning based on Q-learning algorithm
    Li, Shaochuan
    Wang, Xuiqing
    Hu, Liwei
    Liu, Ying
    2019 WORLD ROBOT CONFERENCE SYMPOSIUM ON ADVANCED ROBOTICS AND AUTOMATION (WRC SARA 2019), 2019, : 160 - 165
  • [4] Hexagon-based Q-learning for object search with multiple robots
    Yoon, HU
    Sim, KB
    ADVANCES IN NATURAL COMPUTATION, PT 3, PROCEEDINGS, 2005, 3612 : 713 - 722
  • [5] A Flexible Collision-Free Trajectory Planning for Multiple Robot Arms by Combining Q-Learning and RRT
    Kawabe, Tomoya
    Nishi, Tatsushi
    2022 IEEE 18TH INTERNATIONAL CONFERENCE ON AUTOMATION SCIENCE AND ENGINEERING (CASE), 2022, : 2363 - 2368
  • [6] A path planning approach for mobile robots using short and safe Q-learning
    Du, He
    Hao, Bing
    Zhao, Jianshuo
    Zhang, Jiamin
    Wang, Qi
    Yuan, Qi
    PLOS ONE, 2022, 17 (09):
  • [7] Hybrid Path Planning Algorithm of the Mobile Agent Based on Q-Learning
    Gao, Tengteng
    Li, Caihong
    Liu, Guoming
    Guo, Na
    Wang, Di
    Li, Yongdi
    AUTOMATIC CONTROL AND COMPUTER SCIENCES, 2022, 56 (02) : 130 - 142
  • [8] An optimized Q-Learning algorithm for mobile robot local path planning
    Zhou, Qian
    Lian, Yang
    Wu, Jiayang
    Zhu, Mengyue
    Wang, Haiyong
    Cao, Jinli
    KNOWLEDGE-BASED SYSTEMS, 2024, 286
  • [9] Dynamic Path Planning of a Mobile Robot with Improved Q-Learning algorithm
    Li, Siding
    Xu, Xin
    Zuo, Lei
    2015 IEEE INTERNATIONAL CONFERENCE ON INFORMATION AND AUTOMATION, 2015, : 409 - 414
  • [10] Extended Q-Learning Algorithm for Path-Planning of a Mobile Robot
    Goswami , Indrani
    Das, Pradipta Kumar
    Konar, Amit
    Janarthanan, R.
    SIMULATED EVOLUTION AND LEARNING, 2010, 6457 : 379 - +