Cable SCARA Robot Controlled by a Neural Network Using Reinforcement Learning

被引:0
|
作者
Okabe, Eduardo [1 ]
Paiva, Victor [2 ]
Silva-Teixeira, Luis H. [3 ]
Izuka, Jaime [4 ]
机构
[1] Univ Estadual Campinas, Sch Appl Sci, Rua Pedro Zaccaria 1300, BR-13484350 Limeira, Brazil
[2] Univ Estadual Campinas, Sch Mech Engn, Dept Integrated Syst, Rua Mendeleyev 200, BR-13083860 Campinas, Brazil
[3] Univ Estadual Campinas, Sch Mech Engn, Dept Integrated Syst, R Mendeleyev 200, BR-13083860 Campinas, SP, Brazil
[4] Univ Estadual Campinas, Sch Appl Sci, Rua Pedro Zaccaria 1300, BR-13484350 Limeira, SP, Brazil
来源
JOURNAL OF COMPUTATIONAL AND NONLINEAR DYNAMICS | 2023年 / 18卷 / 10期
关键词
Compendex;
D O I
10.1115/1.4063222
中图分类号
TH [机械、仪表工业];
学科分类号
0802 ;
摘要
In this work, three reinforcement learning algorithms (Proximal Policy Optimization, Soft Actor-Critic, and Twin Delayed Deep Deterministic Policy Gradient) are employed to control a two link selective compliance articulated robot arm (SCARA) robot. This robot has three cables attached to its end-effector, which creates a triangular shaped workspace. Positioning the end-effector in the workspace is a relatively simple kinematic problem, but moving outside this region, although possible, requires a nonlinear dynamic model and a state-of-the-art controller. To solve this problem in a simple manner, reinforcement learning algorithms are used to find possible trajectories for three targets out of the workspace. Additionally, the SCARA mechanism offers two possible configurations for each end-effector position. The algorithm results are compared in terms of displacement error, velocity, and standard deviation among ten trajectories provided by the trained network. The results indicate the Proximal Policy Algorithm as the most consistent in the analyzed situations. Still, the Soft Actor-Critic presented better solutions, and Twin Delayed Deep Deterministic Policy Gradient provided interesting and more unusual trajectories.
引用
收藏
页数:7
相关论文
共 50 条
  • [21] Adaptive neural control using reinforcement learning for a class of robot manipulator
    Li Tang
    Yan-Jun Liu
    Shaocheng Tong
    Neural Computing and Applications, 2014, 25 : 135 - 141
  • [22] Adaptive neural control using reinforcement learning for a class of robot manipulator
    Tang, Li
    Liu, Yan-Jun
    Tong, Shaocheng
    NEURAL COMPUTING & APPLICATIONS, 2014, 25 (01): : 135 - 141
  • [24] Switching Decision of Air-Ground Amphibious Robot using Neural Network-based Reinforcement Learning
    Liu, Zhiyong
    Liu, Yong
    2019 9TH IEEE ANNUAL INTERNATIONAL CONFERENCE ON CYBER TECHNOLOGY IN AUTOMATION, CONTROL, AND INTELLIGENT SYSTEMS (IEEE-CYBER 2019), 2019, : 883 - 888
  • [25] A study on the optimal route design considering time of mobile robot using recurrent neural network and reinforcement learning
    Min Hyuk Woo
    Soo-Hong Lee
    Hye Min Cha
    Journal of Mechanical Science and Technology, 2018, 32 : 4933 - 4939
  • [26] A study on the optimal route design considering time of mobile robot using recurrent neural network and reinforcement learning
    Woo, Min Hyuk
    Lee, Soo-Hong
    Cha, Hye Min
    JOURNAL OF MECHANICAL SCIENCE AND TECHNOLOGY, 2018, 32 (10) : 4933 - 4939
  • [27] Emergence of Prediction by Reinforcement Learning Using a Recurrent Neural Network
    Goto, Kenta
    Shibata, Katsunari
    JOURNAL OF ROBOTICS, 2010, 2010
  • [28] Untying cable by combining 3D deep neural network with deep reinforcement learning
    Fan, Zheming
    Shao, Wanpeng
    Hayashi, Toyohiro
    Ohashi, Takeshi
    ADVANCED ROBOTICS, 2023, 37 (05) : 380 - 394
  • [29] Coarse planning for landmark navigation in a neural-network reinforcement-learning robot
    Baldassarre, G
    IROS 2001: PROCEEDINGS OF THE 2001 IEEE/RJS INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, VOLS 1-4: EXPANDING THE SOCIETAL ROLE OF ROBOTICS IN THE NEXT MILLENNIUM, 2001, : 2398 - 2403
  • [30] Neural Network Ensembles in Reinforcement Learning
    Stefan Faußer
    Friedhelm Schwenker
    Neural Processing Letters, 2015, 41 : 55 - 69