Learning latent representations to co-adapt to humans

被引:1
|
作者
Parekh, Sagar [1 ]
Losey, Dylan P. [1 ]
机构
[1] Virginia Tech, Mech Engn Dept, Blacksburg, VA 24060 USA
基金
美国食品与农业研究所;
关键词
Human-robot interaction; Representation learning; Reinforcement learning; ROBOT; GAMES;
D O I
10.1007/s10514-023-10109-5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
When robots interact with humans in homes, roads, or factories the human's behavior often changes in response to the robot. Non-stationary humans are challenging for robot learners: actions the robot has learned to coordinate with the original human may fail after the human adapts to the robot. In this paper we introduce an algorithmic formalism that enables robots (i.e., ego agents) to co-adapt alongside dynamic humans (i.e., other agents) using only the robot's low-level states, actions, and rewards. A core challenge is that humans not only react to the robot's behavior, but the way in which humans react inevitably changes both over time and between users. To deal with this challenge, our insight is that-instead of building an exact model of the human-robots can learn and reason over high-level representations of the human's policy and policy dynamics. Applying this insight we develop RILI: Robustly Influencing Latent Intent. RILI first embeds low-level robot observations into predictions of the human's latent strategy and strategy dynamics. Next, RILI harnesses these predictions to select actions that influence the adaptive human towards advantageous, high reward behaviors over repeated interactions. We demonstrate that-given RILI's measured performance with users sampled from an underlying distribution-we can probabilistically bound RILI's expected performance across new humans sampled from the same distribution. Our simulated experiments compare RILI to state-of-the-art representation and reinforcement learning baselines, and show that RILI better learns to coordinate with imperfect, noisy, and time-varying agents. Finally, we conduct two user studies where RILI co-adapts alongside actual humans in a game of tag and a tower-building task. See videos of our user studies here: https://youtu.be/WYGO5amDXbQ
引用
下载
收藏
页码:771 / 796
页数:26
相关论文
共 50 条
  • [31] Unsupervised learning reveals interpretable latent representations for translucency perception
    Liao, Chenxi W.
    Sawayama, Masataka
    Xiao, Bei W.
    PLOS COMPUTATIONAL BIOLOGY, 2023, 19 (02)
  • [32] Learning Time Series Counterfactuals via Latent Space Representations
    Wang, Zhendong
    Samsten, Isak
    Mochaourab, Rami
    Papapetrou, Panagiotis
    DISCOVERY SCIENCE (DS 2021), 2021, 12986 : 369 - 384
  • [33] FoLaR: Foggy Latent Representations for Reinforcement Learning with Partial Observability
    Meisheri, Hardik
    Khadilkar, Harshad
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [34] Leveraging maximum entropy and correlation on latent factors for learning representations
    He, Zhicheng
    Liu, Jie
    Dang, Kai
    Zhuang, Fuzhen
    Huang, Yalou
    Neural Networks, 2020, 131 : 312 - 323
  • [35] Co-active Learning to Adapt Humanoid Movement for Manipulation
    Mao, Ren
    Baras, John S.
    Yang, Yezhou
    Fermueller, Cornelia
    2016 IEEE-RAS 16TH INTERNATIONAL CONFERENCE ON HUMANOID ROBOTS (HUMANOIDS), 2016, : 372 - 378
  • [37] Learning to Adapt
    Guo, Yi
    IEEE ROBOTICS & AUTOMATION MAGAZINE, 2021, 28 (04) : 4 - 6
  • [38] Learning to adapt
    不详
    NATURE CLIMATE CHANGE, 2011, 1 (05) : 274 - 274
  • [39] Learning to adapt
    Stokstad, E
    SCIENCE, 2005, 309 (5735) : 688 - 690
  • [40] Associative learning and latent inhibition in a conditioned suppression paradigm in humans
    Salgado, JV
    Vidal, M
    Oberling, P
    Graeff, FG
    Danion, JM
    Sandner, G
    BEHAVIOURAL BRAIN RESEARCH, 2000, 117 (1-2) : 53 - 60