Deep Reinforcement Learning with Inverse Jacobian based Model-Free Path Planning for Deburring in Complex Industrial Environment

被引:2
|
作者
Rahul, M. R. [1 ]
Chiddarwar, Shital S. [1 ]
机构
[1] VNIT, Dept Mech Engn, Nagpur, Maharashtra, India
关键词
Path planning; Deburring; TD3; Deep reinforcement learning; Collision avoidance; ROADMAP; MOTION;
D O I
10.1007/s10846-023-02030-x
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this study, we present an innovative approach to robotic deburring path planning by combining deep reinforcement learning (DRL) with an inverse Jacobian strategy. Existing model-based path planning methods, including sampling-based approaches, often suffer from computational complexity and challenges in capturing the dynamics of deburring systems. To overcome these limitations, our novel DRL-based framework for path planning leverages experiential learning to identify optimal deburring trajectories without relying on predefined models. This model-free approach is particularly suited for complex deburring scenarios with unknown system dynamics. Additionally, we employ an inverse Jacobian technique with a time-varying gain module (eta(t) = e<^>2t) during training, which yields remarkable benefits in terms of exploration-exploitation balance and collision avoidance, enhancing the overall performance of the DRL agent. Through a series of experiments conducted in a simulated environment, we evaluate the efficacy of our proposed algorithm for deburring path planning. Our modified DRL-based approach, utilizing inverse kinematics with a time-varying gain module, demonstrates superior performance in terms of convergence speed, optimality, and robustness when compared to conventional path planning methods. Notably, in comparison to algorithms like sampling-based strategies, our model-free DRL-based approach outperforms these methods, achieving an exceptional average success rate of 97%. The integration of the inverse Jacobian technique further enhances the effectiveness of our algorithm by effectively reducing the state space dimensionality, leading to improved learning efficiency and the generation of optimal deburring trajectories.
引用
收藏
页数:19
相关论文
共 50 条
  • [1] Deep Reinforcement Learning with Inverse Jacobian based Model-Free Path Planning for Deburring in Complex Industrial Environment
    M. R. Rahul
    Shital S. Chiddarwar
    [J]. Journal of Intelligent & Robotic Systems, 2024, 110
  • [2] A Model-free Deep Reinforcement Learning Approach for Robotic Manipulators Path Planning
    Liu, Wenxing
    Niu, Hanlin
    Mahyuddin, Muhammad Nasiruddin
    Herrmann, Guido
    Carrasco, Joaquin
    [J]. 2021 21ST INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND SYSTEMS (ICCAS 2021), 2021, : 512 - 517
  • [3] Optimization of Smart Textiles Robotic Arm Path Planning: A Model-Free Deep Reinforcement Learning Approach with Inverse Kinematics
    Zhao, Di
    Ding, Zhenyu
    Li, Wenjie
    Zhao, Sen
    Du, Yuhong
    [J]. PROCESSES, 2024, 12 (01)
  • [4] Model-Free Deep Inverse Reinforcement Learning by Logistic Regression
    Eiji Uchibe
    [J]. Neural Processing Letters, 2018, 47 : 891 - 905
  • [6] A path planning method based on deep reinforcement learning for AUV in complex marine environment
    Zhang, An
    Wang, Weixiang
    Bi, Wenhao
    Huang, Zhanjun
    [J]. Ocean Engineering, 2024, 313
  • [7] Dynamic Path Planning of Unknown Environment Based on Deep Reinforcement Learning
    Lei, Xiaoyun
    Zhang, Zhian
    Dong, Peifang
    [J]. JOURNAL OF ROBOTICS, 2018, 2018
  • [8] A decentralized path planning model based on deep reinforcement learning
    Guo, Dong
    Ji, Shouwen
    Yao, Yanke
    Chen, Cheng
    [J]. COMPUTERS & ELECTRICAL ENGINEERING, 2024, 117
  • [9] Path planning in an unknown environment based on deep reinforcement learning with prior knowledge
    Lou, Ping
    Xu, Kun
    Jiang, Xuemei
    Xiao, Zheng
    Yan, Junwei
    [J]. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2021, 41 (06) : 5773 - 5789
  • [10] Robot path planning based on deep reinforcement learning
    Long, Yinxin
    He, Huajin
    [J]. 2020 IEEE CONFERENCE ON TELECOMMUNICATIONS, OPTICS AND COMPUTER SCIENCE (TOCS), 2020, : 151 - 154