Improved Q-Learning Method for Multirobot Formation and Path Planning with Concave Obstacles

被引:1
|
作者
Fan, Zhilin [1 ]
Liu, Fei [1 ]
Ning, Xinshun [1 ]
Han, Yilin [1 ]
Wang, Jian [2 ]
Yang, Hongyong [1 ]
Liu, Li [1 ]
机构
[1] Ludong Univ, Sch Informat & Elect Engn, Yantai 264000, Peoples R China
[2] Yantai Municipal Peoples Procuratorate, Yantai 264000, Peoples R China
关键词
MOBILE; ALGORITHM;
D O I
10.1155/2021/4294841
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Aiming at the formation and path planning of multirobot systems in an unknown environment, a path planning method for multirobot formation based on improved Q-learning is proposed. Based on the leader-following approach, the leader robot uses an improved Q-learning algorithm to plan the path and the follower robot achieves a tracking strategy of gravitational potential field (GPF) by designing a cost function to select actions. Specifically, to improve the Q-learning, Q-value is initialized by environmental guidance of the target's GPF. Then, the virtual obstacle-filling avoidance strategy is presented to fill non-obstacles which is judged to tend to concave obstacles with virtual obstacles. Besides, the simulated annealing (SA) algorithm whose controlling temperature is adjusted in real time according to the learning situation of the Q-learning is applied to improve the action selection strategy. The experimental results show that the improved Q-learning algorithm reduces the convergence time by 89.9% and the number of convergence rounds by 63.4% compared with the traditional algorithm. With the help of the method, multiple robots have a clear division of labor and quickly plan a globally optimized formation path in a completely unknown environment.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] ETQ-learning: an improved Q-learning algorithm for path planning
    Wang, Huanwei
    Jing, Jing
    Wang, Qianlv
    He, Hongqi
    Qi, Xuyan
    Lou, Rui
    [J]. INTELLIGENT SERVICE ROBOTICS, 2024, 17 (04) : 915 - 929
  • [2] A Path Planning Algorithm for UAV Based on Improved Q-Learning
    Yan, Chao
    Xiang, Xiaojia
    [J]. 2018 2ND INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION SCIENCES (ICRAS), 2018, : 46 - 50
  • [3] Ship Local Path Planning Based on Improved Q-Learning
    Gong, Ming-Fan
    Xu, Hai-Xiang
    Feng, Hui
    Wang, Yong
    Xue, Xue-Hua
    [J]. Chuan Bo Li Xue/Journal of Ship Mechanics, 2022, 26 (06): : 824 - 833
  • [4] A Deterministic Improved Q-Learning for Path Planning of a Mobile Robot
    Konar, Amit
    Chakraborty, Indrani Goswami
    Singh, Sapam Jitu
    Jain, Lakhmi C.
    Nagar, Atulya K.
    [J]. IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2013, 43 (05): : 1141 - 1153
  • [5] The Method Based on Q-Learning Path Planning in Migrating Workflow
    Xiao, Song
    Wang, Xiao-lin
    [J]. PROCEEDINGS 2013 INTERNATIONAL CONFERENCE ON MECHATRONIC SCIENCES, ELECTRIC ENGINEERING AND COMPUTER (MEC), 2013, : 2204 - 2208
  • [6] Dynamic Path Planning of a Mobile Robot with Improved Q-Learning algorithm
    Li, Siding
    Xu, Xin
    Zuo, Lei
    [J]. 2015 IEEE INTERNATIONAL CONFERENCE ON INFORMATION AND AUTOMATION, 2015, : 409 - 414
  • [7] PATH PLANNING OF MOBILE ROBOT BASED ON THE IMPROVED Q-LEARNING ALGORITHM
    Chen, Chaorui
    Wang, Dongshu
    [J]. INTERNATIONAL JOURNAL OF INNOVATIVE COMPUTING INFORMATION AND CONTROL, 2022, 18 (03): : 687 - 702
  • [8] Improved Q-Learning Applied to Dynamic Obstacle Avoidance and Path Planning
    Wang, Chunlei
    Yang, Xiao
    Li, He
    [J]. IEEE ACCESS, 2022, 10 : 92879 - 92888
  • [9] A Dynamic Hidden Forwarding Path Planning Method Based on Improved Q-Learning in SDN Environments
    Chen, Yun
    Lv, Kun
    Hu, Changzhen
    [J]. SECURITY AND COMMUNICATION NETWORKS, 2018,
  • [10] Path planning of mobile robots with Q-learning
    Cetin, Halil
    Durdu, Akif
    [J]. 2014 22ND SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2014, : 2162 - 2165