Table-Balancing Cooperative Robot Based on Deep Reinforcement Learning

被引:1
|
作者
Kim, Yewon [1 ]
Kim, Dae-Won [2 ]
Kang, Bo-Yeong [3 ]
机构
[1] Kyungpook Natl Univ, Dept Artificial Intelligence, Daegu 41566, South Korea
[2] Chung Ang Univ, Sch Comp Sci & Engn, 84 Heukseok Ro, Seoul 06974, South Korea
[3] Kyungpook Natl Univ, Dept Robot & Smart Syst Engn, Daegu 41566, South Korea
基金
新加坡国家研究基金会;
关键词
reinforcement learning; deep Q-network; cooperative robot; human-robot interaction; FRAMEWORK;
D O I
10.3390/s23115235
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
Reinforcement learning is one of the artificial intelligence methods that enable robots to judge and operate situations on their own by learning to perform tasks. Previous reinforcement learning research has mainly focused on tasks performed by individual robots; however, everyday tasks, such as balancing tables, often require cooperation between two individuals to avoid injury when moving. In this research, we propose a deep reinforcement learning-based technique for robots to perform a table-balancing task in cooperation with a human. The cooperative robot proposed in this paper recognizes human behavior to balance the table. This recognition is achieved by utilizing the robot's camera to take an image of the state of the table, then the table-balance action is performed afterward. Deep Q-network (DQN) is a deep reinforcement learning technology applied to cooperative robots. As a result of learning table balancing, on average, the cooperative robot showed a 90% optimal policy convergence rate in 20 runs of training with optimal hyperparameters applied to DQN-based techniques. In the H/W experiment, the trained DQN-based robot achieved an operation precision of 90%, thus verifying its excellent performance.
引用
收藏
页数:16
相关论文
共 50 条
  • [11] Cooperative Multi-Robot Navigation in Dynamic Environment with Deep Reinforcement Learning
    Han, Ruihua
    Chen, Shengduo
    Hao, Qi
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 448 - 454
  • [12] Learning to Navigate in Turbulent Flows With Aerial Robot Swarms: A Cooperative Deep Reinforcement Learning Approach
    Patino, Diego
    Mayya, Siddharth
    Calderon, Juan
    Daniilidis, Kostas
    Saldana, David
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (07) : 4219 - 4226
  • [13] Multi-robot cooperative behavior generation based on reinforcement learning
    Li, Dong-Mei
    Chen, Wei-Dong
    Xi, Yu-Geng
    [J]. Shanghai Jiaotong Daxue Xuebao/Journal of Shanghai Jiaotong University, 2005, 39 (08): : 1331 - 1335
  • [14] Cooperative behavior of a heterogeneous robot team for planetary exploration using deep reinforcement learning
    Barth, Andrew
    Ma, Ou
    [J]. ACTA ASTRONAUTICA, 2024, 214 : 689 - 700
  • [15] Cooperative behavior of a heterogeneous robot team for planetary exploration using deep reinforcement learning
    Barth, Andrew
    Ma, Ou
    [J]. Acta Astronautica, 2024, 214 : 689 - 700
  • [16] Deep Reinforcement Learning Based Mobile Robot Navigation: A Review
    Zhu, Kai
    Zhang, Tao
    [J]. TSINGHUA SCIENCE AND TECHNOLOGY, 2021, 26 (05) : 674 - 691
  • [17] Towards Deep Reinforcement Learning based Chinese Calligraphy Robot
    Wu, Ruiqi
    Fang, Wubing
    Chao, Fei
    Gao, Xingen
    Zhou, Changle
    Yang, Longzhi
    Lin, Chih-Min
    Shang, Changjing
    [J]. 2018 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS (ROBIO), 2018, : 507 - 512
  • [18] Robot Obstacle Avoidance Controller Based on Deep Reinforcement Learning
    Tang, Yaokun
    Chen, Qingyu
    Wei, Yuxin
    [J]. JOURNAL OF SENSORS, 2022, 2022
  • [19] Robot Navigation with Interaction-based Deep Reinforcement Learning
    Zhai, Yu
    Miao, Yanzi
    Wang, Hesheng
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS (IEEE-ROBIO 2021), 2021, : 1974 - 1979
  • [20] Research on path planning of robot based on deep reinforcement learning
    Liu, Feng
    Chen, Chang
    Li, Zhihua
    Guan, Zhi-Hong
    Wang, Hua O.
    [J]. PROCEEDINGS OF THE 39TH CHINESE CONTROL CONFERENCE, 2020, : 3730 - 3734