Real-World Human-Robot Collaborative Reinforcement Learning

被引:9
|
作者
Shafti, Ali [1 ,2 ]
Tjomsland, Jonas [1 ,2 ]
Dudley, William [1 ,2 ]
Faisal, A. Aldo [1 ,2 ]
机构
[1] Imperial Coll London, Dept Bioengn, Brain & Behav Lab, London SW7 2AZ, England
[2] Imperial Coll London, Dept Comp, London SW7 2AZ, England
关键词
D O I
10.1109/IROS45743.2020.9341473
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The intuitive collaboration of humans and intelligent robots (embodied AI) in the real-world is an essential objective for many desirable applications of robotics. Whilst there is much research regarding explicit communication, we focus on how humans and robots interact implicitly, on motor adaptation level. We present a real-world setup of a human-robot collaborative maze game, designed to be non-trivial and only solvable through collaboration, by limiting the actions to rotations of two orthogonal axes, and assigning each axes to one player. This results in neither the human nor the agent being able to solve the game on their own. We use deep reinforcement learning for the control of the robotic agent, and achieve results within 30 minutes of real-world play, without any type of pre-training. We then use this setup to perform systematic experiments on human/agent behaviour and adaptation when co-learning a policy for the collaborative game. We present results on how co-policy learning occurs over time between the human and the robotic agent resulting in each participant's agent serving as a representation of how they would play the game. This allows us to relate a person's success when playing with different agents than their own, by comparing the policy of the agent with that of their own agent.
引用
收藏
页码:11161 / 11166
页数:6
相关论文
共 50 条
  • [1] Intrinsically motivated reinforcement learning for human-robot interaction in the real-world
    Qureshi, Ahmed Hussain
    Nakamura, Yutaka
    Yoshikawa, Yuichiro
    Ishiguro, Hiroshi
    [J]. NEURAL NETWORKS, 2018, 107 : 23 - 33
  • [2] A Human-Robot Collaborative Reinforcement Learning Algorithm
    Uri Kartoun
    Helman Stern
    Yael Edan
    [J]. Journal of Intelligent & Robotic Systems, 2010, 60 : 217 - 239
  • [3] A Human-Robot Collaborative Reinforcement Learning Algorithm
    Kartoun, Uri
    Stern, Helman
    Edan, Yael
    [J]. JOURNAL OF INTELLIGENT & ROBOTIC SYSTEMS, 2010, 60 (02) : 217 - 239
  • [4] Embodied Affect for Real-World Human-Robot Interaction
    Canamero, Lola
    [J]. PROCEEDINGS OF THE 2020 ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION (HRI '20), 2020, : 459 - 459
  • [5] Real-world oriented distributed human-robot interface system
    Iwakura, Y
    Shiraishi, Y
    Nakauchi, Y
    Anzai, Y
    [J]. RO-MAN '97 SENDAI: 6TH IEEE INTERNATIONAL WORKSHOP ON ROBOT AND HUMAN COMMUNICATION, PROCEEDINGS, 1997, : 188 - 193
  • [6] Setting up a Reinforcement Learning Task with a Real-World Robot
    Mahmood, A. Rupam
    Korenkevych, Dmytro
    Komer, Brent J.
    Bergstra, James
    [J]. 2018 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2018, : 4635 - 4640
  • [7] Real-world reinforcement learning for autonomous humanoid robot docking
    Navarro-Guerrero, Nicolas
    Weber, Cornelius
    Schroeter, Pascal
    Wermter, Stefan
    [J]. ROBOTICS AND AUTONOMOUS SYSTEMS, 2012, 60 (11) : 1400 - 1407
  • [8] Mastering the Working Sequence in Human-Robot Collaborative Assembly Based on Reinforcement Learning
    Yu, Tian
    Huang, Jing
    Chang, Qing
    [J]. IEEE ACCESS, 2020, 8 : 163868 - 163877
  • [9] Real-world Robot Reaching Skill Learning Based on Deep Reinforcement Learning
    Liu, Naijun
    Lu, Tao
    Cai, Yinghao
    Wang, Rui
    Wang, Shuo
    [J]. PROCEEDINGS OF THE 32ND 2020 CHINESE CONTROL AND DECISION CONFERENCE (CCDC 2020), 2020, : 4780 - 4784
  • [10] Robust Assembly Sequence Generation in a Human-Robot Collaborative Workcell by Reinforcement Learning
    Antonelli, Dario
    Zeng, Qingfei
    Aliev, Khurshid
    Liu, Xuemei
    [J]. FME TRANSACTIONS, 2021, 49 (04): : 851 - 858