An actor-critic learning framework based on Lyapunov stability for automatic assembly

被引:0
|
作者
Xinwang Li
Juliang Xiao
Yu Cheng
Haitao Liu
机构
[1] Key Laboratory of Mechanism Theory and Equipment Design of Ministry of Education Tianjin University,
[2] AVIC Manufacturing Technology Institute,undefined
来源
Applied Intelligence | 2023年 / 53卷
关键词
Reinforcement learning; Automatic assembly; Lyapunov stability; Reward reshaping; Median Q-value;
D O I
暂无
中图分类号
学科分类号
摘要
With the continuous improvement of the reinforcement learning (RL) algorithm, the algorithm has achieved excellent performance in an increasing number of automatic control tasks. However, there are still some challenges when applying the algorithm to realistic automatic assembly. The most significant challenge is that the stability of these model-free RL methods cannot be effectively guaranteed. Stability is the most critical characteristic of a control system, and stability is closely related to reliability and safety. To ensure the stability of the system, we reconstruct the RL algorithm based on the Lyapunov stability theory of the stochastic system proposed in this paper. An actor-critic learning framework based on Lyapunov stability (LSAC) is proposed for automatic assembly. In addition, this paper proposes a median Q-value theory to alleviate the Q-value estimation deviation that restricts the performance of the RL algorithm. To allow RL agents to better complete the automatic assembly task, this paper designs an adaptive impedance control algorithm. This impedance algorithm executes the actions output by the LSAC framework. Finally, a realistic experiment on automatic assembly is carried out to verify the robustness and superiority of the proposed strategy.
引用
收藏
页码:4801 / 4812
页数:11
相关论文
共 50 条
  • [1] An actor-critic learning framework based on Lyapunov stability for automatic assembly
    Li, Xinwang
    Xiao, Juliang
    Cheng, Yu
    Liu, Haitao
    [J]. APPLIED INTELLIGENCE, 2023, 53 (04) : 4801 - 4812
  • [2] Actor-Critic Reinforcement Learning for Control With Stability Guarantee
    Han, Minghao
    Zhang, Lixian
    Wang, Jun
    Pan, Wei
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2020, 5 (04) : 6217 - 6224
  • [3] Actor-Critic based Improper Reinforcement Learning
    Zaki, Mohammadi
    Mohan, Avinash
    Gopalan, Aditya
    Mannor, Shie
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [4] An Actor-Critic Framework for Online Control With Environment Stability Guarantee
    Osinenko, Pavel
    Yaremenko, Grigory
    Malaniya, Georgiy
    Bolychev, Anton
    [J]. IEEE ACCESS, 2023, 11 : 89188 - 89204
  • [5] QVDDPG: QV Learning with Balanced Constraint in Actor-Critic Framework
    Huang, Jiao
    Hu, Jifeng
    Yang, Luheng
    Ren, Zhihang
    Chen, Hechang
    Yang, Bo
    [J]. 2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [6] Actor-Critic Learning Based on Adaptive Importance Sampling
    Cheng Yuhu
    Feng Huanting
    Wang Xuesong
    [J]. CHINESE JOURNAL OF ELECTRONICS, 2010, 19 (04) : 583 - 588
  • [7] Discrete-Time Lyapunov based Kinematic Control of Robot Manipulator using Actor-Critic Framework
    Kamboj, Ankur
    Prakash, Ravi
    Mohanta, Jayant Kumar
    Behera, Laxmidhar
    [J]. 2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [8] Actor-critic learning based on fuzzy inference system
    Jouffe, L
    [J]. INFORMATION INTELLIGENCE AND SYSTEMS, VOLS 1-4, 1996, : 339 - 344
  • [9] Actor-Critic Reinforcement Learning for Automatic Left Atrial Appendage Segmentation
    Abdullah, Al Walid
    Yun, Il Dong
    [J]. PROCEEDINGS 2018 IEEE INTERNATIONAL CONFERENCE ON BIOINFORMATICS AND BIOMEDICINE (BIBM), 2018, : 609 - 612
  • [10] DTaxa: An actor-critic for automatic taxonomy induction
    Han, Yongming
    Lang, Yanwei
    Cheng, Minjie
    Geng, Zhiqiang
    Chen, Guofei
    Xia, Tao
    [J]. ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2021, 106