An actor-critic learning framework based on Lyapunov stability for automatic assembly

被引:1
|
作者
Li, Xinwang [1 ,2 ]
Xiao, Juliang [1 ]
Cheng, Yu [1 ]
Liu, Haitao [1 ]
机构
[1] Tianjin Univ, Key Lab Mech Theory & Equipment Design, Minist Educ, Tianjin 300354, Peoples R China
[2] AVIC Mfg Technol Inst, Beijing 100024, Peoples R China
基金
中国国家自然科学基金;
关键词
Reinforcement learning; Automatic assembly; Lyapunov stability; Reward reshaping; Median Q-value;
D O I
10.1007/s10489-022-03844-2
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the continuous improvement of the reinforcement learning (RL) algorithm, the algorithm has achieved excellent performance in an increasing number of automatic control tasks. However, there are still some challenges when applying the algorithm to realistic automatic assembly. The most significant challenge is that the stability of these model-free RL methods cannot be effectively guaranteed. Stability is the most critical characteristic of a control system, and stability is closely related to reliability and safety. To ensure the stability of the system, we reconstruct the RL algorithm based on the Lyapunov stability theory of the stochastic system proposed in this paper. An actor-critic learning framework based on Lyapunov stability (LSAC) is proposed for automatic assembly. In addition, this paper proposes a median Q-value theory to alleviate the Q-value estimation deviation that restricts the performance of the RL algorithm. To allow RL agents to better complete the automatic assembly task, this paper designs an adaptive impedance control algorithm. This impedance algorithm executes the actions output by the LSAC framework. Finally, a realistic experiment on automatic assembly is carried out to verify the robustness and superiority of the proposed strategy.
引用
收藏
页码:4801 / 4812
页数:12
相关论文
共 50 条
  • [1] An actor-critic learning framework based on Lyapunov stability for automatic assembly
    Xinwang Li
    Juliang Xiao
    Yu Cheng
    Haitao Liu
    [J]. Applied Intelligence, 2023, 53 : 4801 - 4812
  • [2] Actor-Critic Reinforcement Learning for Control With Stability Guarantee
    Han, Minghao
    Zhang, Lixian
    Wang, Jun
    Pan, Wei
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2020, 5 (04) : 6217 - 6224
  • [3] Actor-Critic based Improper Reinforcement Learning
    Zaki, Mohammadi
    Mohan, Avinash
    Gopalan, Aditya
    Mannor, Shie
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [4] An Actor-Critic Framework for Online Control With Environment Stability Guarantee
    Osinenko, Pavel
    Yaremenko, Grigory
    Malaniya, Georgiy
    Bolychev, Anton
    [J]. IEEE ACCESS, 2023, 11 : 89188 - 89204
  • [5] QVDDPG: QV Learning with Balanced Constraint in Actor-Critic Framework
    Huang, Jiao
    Hu, Jifeng
    Yang, Luheng
    Ren, Zhihang
    Chen, Hechang
    Yang, Bo
    [J]. 2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [6] Actor-Critic Learning Based on Adaptive Importance Sampling
    Cheng Yuhu
    Feng Huanting
    Wang Xuesong
    [J]. CHINESE JOURNAL OF ELECTRONICS, 2010, 19 (04) : 583 - 588
  • [7] Discrete-Time Lyapunov based Kinematic Control of Robot Manipulator using Actor-Critic Framework
    Kamboj, Ankur
    Prakash, Ravi
    Mohanta, Jayant Kumar
    Behera, Laxmidhar
    [J]. 2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [8] Actor-critic learning based on fuzzy inference system
    Jouffe, L
    [J]. INFORMATION INTELLIGENCE AND SYSTEMS, VOLS 1-4, 1996, : 339 - 344
  • [9] Actor-Critic Reinforcement Learning for Automatic Left Atrial Appendage Segmentation
    Abdullah, Al Walid
    Yun, Il Dong
    [J]. PROCEEDINGS 2018 IEEE INTERNATIONAL CONFERENCE ON BIOINFORMATICS AND BIOMEDICINE (BIBM), 2018, : 609 - 612
  • [10] DTaxa: An actor-critic for automatic taxonomy induction
    Han, Yongming
    Lang, Yanwei
    Cheng, Minjie
    Geng, Zhiqiang
    Chen, Guofei
    Xia, Tao
    [J]. ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2021, 106