A Path Planning Algorithm for Space Manipulator Based on Q-Learning

被引:0
|
作者
Li, Taiguo [1 ]
Li, Quanhong [2 ]
Li, Wenxi [1 ]
Xia, Jiagao [1 ]
Tang, Wenhua [1 ]
Wang, Weiwen [1 ]
机构
[1] Lanzhou Inst Phys, Lanzhou, Gansu, Peoples R China
[2] Gansu Agr Univ, Coll Resources & Environm, Lanzhou, Gansu, Peoples R China
关键词
Space Manipulato; Grid Model; Q-Learning; Reinforcement Learning; Path Planning;
D O I
10.1109/itaic.2019.8785427
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
an improved Q-Learning autonomous learning algorithm is proposed to solve the problem of the adaptive path planning of the space manipulator in the unknown environment. After simplification of the manipulator and obstacle model, the grid model of the environment is established, and the position of the manipulator and obstacles are randomly deployed in the grid map. Based on the analysis of the basic principle of reinforcement learning and the state generalization method, the improved Q-Learning algorithm is used to carry out the path planning. In this algorithm, the reward and punishment strategies in the path planning of the manipulator are designed, and the approximate greedy and continuous micro Botlzmann distribution behavior selection strategy is adopted. According to the autonomous learning of Q-table, the manipulator can guide its follow-up action selection and path planning, reduce the number of manipulator movement, and reduce the blindness of the learning process. The results show that the algorithm has the advantages of simple calculation, strong self-learning ability, and can successfully complete the adaptive path planning in unknown environment.
引用
下载
收藏
页码:1566 / 1571
页数:6
相关论文
共 50 条
  • [41] An immune plasma algorithm with Q-learning based pandemic management for path planning of unmanned aerial vehicles
    Aslan, Selcuk
    Demirci, Sercan
    EGYPTIAN INFORMATICS JOURNAL, 2024, 26
  • [42] Path planning of a mobile robot in a free-space environment using Q-learning
    Jiang, Jianxun
    Xin, Jianbin
    PROGRESS IN ARTIFICIAL INTELLIGENCE, 2019, 8 (01) : 133 - 142
  • [43] Path planning of a mobile robot in a free-space environment using Q-learning
    Jianxun Jiang
    Jianbin Xin
    Progress in Artificial Intelligence, 2019, 8 : 133 - 142
  • [44] Improved Q-Learning Algorithm Based on Flower Pollination Algorithm and Tabulation Method for Unmanned Aerial Vehicle Path Planning
    Bo, Lan
    Zhang, Tiezhu
    Zhang, Hongxin
    Yang, Jian
    Zhang, Zhen
    Zhang, Caihong
    Liu, Mingjie
    IEEE ACCESS, 2024, 12 : 104429 - 104444
  • [45] The Experience-Memory Q-Learning Algorithm for Robot Path Planning in Unknown Environment
    Zhao, Meng
    Lu, Hui
    Yang, Siyi
    Guo, Fengjuan
    IEEE ACCESS, 2020, 8 : 47824 - 47844
  • [46] A modified Q-learning algorithm for robot path planning in a digital twin assembly system
    Guo, Xiaowei
    Peng, Gongzhuang
    Meng, Yingying
    INTERNATIONAL JOURNAL OF ADVANCED MANUFACTURING TECHNOLOGY, 2022, 119 (5-6): : 3951 - 3961
  • [47] A modified Q-learning algorithm for robot path planning in a digital twin assembly system
    Xiaowei Guo
    Gongzhuang Peng
    Yingying Meng
    The International Journal of Advanced Manufacturing Technology, 2022, 119 : 3951 - 3961
  • [48] Path planning algorithm of space manipulator based on chaos particle swarm optimization algorithm
    Xia, Hong-Wei
    Zhai, Yan-Bin
    Ma, Guang-Cheng
    Deng, Ya
    Wang, Chang-Hong
    Zhongguo Guanxing Jishu Xuebao/Journal of Chinese Inertial Technology, 2014, 22 (02): : 211 - 216
  • [49] Personalized Optimal Bicycle Trip Planning Based on Q-learning Algorithm
    Chen, Yun
    Yan, Wen
    Li, Chunguo
    Huang, Yongming
    Yang, Luxi
    2018 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE (WCNC), 2018,
  • [50] Q-learning based Path Planning Method for UAVs using Priority Shifting
    de Carvalho, Kevin B.
    de Oliveira, Iure Rosa L.
    Villa, Daniel K. D.
    Caldeira, Alexandre G.
    Sarcinelli-Filho, Mario
    Brandao, Alexandre S.
    2022 INTERNATIONAL CONFERENCE ON UNMANNED AIRCRAFT SYSTEMS (ICUAS), 2022, : 421 - 426