Learning latent representations to co-adapt to humans

被引:1
|
作者
Parekh, Sagar [1 ]
Losey, Dylan P. [1 ]
机构
[1] Virginia Tech, Mech Engn Dept, Blacksburg, VA 24060 USA
基金
美国食品与农业研究所;
关键词
Human-robot interaction; Representation learning; Reinforcement learning; ROBOT; GAMES;
D O I
10.1007/s10514-023-10109-5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
When robots interact with humans in homes, roads, or factories the human's behavior often changes in response to the robot. Non-stationary humans are challenging for robot learners: actions the robot has learned to coordinate with the original human may fail after the human adapts to the robot. In this paper we introduce an algorithmic formalism that enables robots (i.e., ego agents) to co-adapt alongside dynamic humans (i.e., other agents) using only the robot's low-level states, actions, and rewards. A core challenge is that humans not only react to the robot's behavior, but the way in which humans react inevitably changes both over time and between users. To deal with this challenge, our insight is that-instead of building an exact model of the human-robots can learn and reason over high-level representations of the human's policy and policy dynamics. Applying this insight we develop RILI: Robustly Influencing Latent Intent. RILI first embeds low-level robot observations into predictions of the human's latent strategy and strategy dynamics. Next, RILI harnesses these predictions to select actions that influence the adaptive human towards advantageous, high reward behaviors over repeated interactions. We demonstrate that-given RILI's measured performance with users sampled from an underlying distribution-we can probabilistically bound RILI's expected performance across new humans sampled from the same distribution. Our simulated experiments compare RILI to state-of-the-art representation and reinforcement learning baselines, and show that RILI better learns to coordinate with imperfect, noisy, and time-varying agents. Finally, we conduct two user studies where RILI co-adapts alongside actual humans in a game of tag and a tower-building task. See videos of our user studies here: https://youtu.be/WYGO5amDXbQ
引用
下载
收藏
页码:771 / 796
页数:26
相关论文
共 50 条
  • [41] Learning Modality-Invariant Latent Representations for Generalized Zero-shot Learning
    Li, Jingjing
    Jing, Mengmeng
    Zhu, Lei
    Ding, Zhengming
    Lu, Ke
    Yang, Yang
    MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, 2020, : 1348 - 1356
  • [42] Learning Human Action Recognition Representations Without Real Humans
    Zhong, Howard
    Mishra, Samarth
    Kim, Donghyun
    Jin, SouYoung
    Panda, Rameswar
    Kuehne, Hilde
    Karlinsky, Leonid
    Saligrama, Venkatesh
    Oliva, Aude
    Feris, Rogerio
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [43] HOW HUMANS ADAPT - ORTNER,DJ
    STEEGMANN, AT
    AMERICAN JOURNAL OF PHYSICAL ANTHROPOLOGY, 1984, 65 (03) : 333 - 334
  • [44] Private-Shared Disentangled Multimodal VAE for Learning of Latent Representations
    Lee, Mihee
    Pavlovic, Vladimir
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2021, 2021, : 1692 - 1700
  • [45] Learning and visualizing chronic latent representations using electronic health records
    David Chushig-Muzo
    Cristina Soguero-Ruiz
    Pablo de Miguel Bohoyo
    Inmaculada Mora-Jiménez
    BioData Mining, 15
  • [46] Learning montages of transformed latent images as representations of objects that change in appearance
    Pal, C
    Frey, BJ
    Jojic, N
    COMPUTER VISION - ECCV 2002, PT IV, 2002, 2353 : 715 - 731
  • [47] Learning Compact Hyperbolic Representations of Latent Space for Old Photo Restoration
    Chen, Rui
    Guo, Tao
    Mu, Yang
    Shen, Li
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 3578 - 3589
  • [48] Learning and visualizing chronic latent representations using electronic health records
    Chushig-Muzo, David
    Soguero-Ruiz, Cristina
    Bohoyo, Pablo de Miguel
    Mora-Jimenez, Inmaculada
    BIODATA MINING, 2022, 15 (01)
  • [49] Learning Generalizable Latent Representations for Novel Degradations in Super-Resolution
    Li, Fengjun
    Feng, Xin
    Chen, Fanglin
    Lu, Guangming
    Pei, Wenjie
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 1797 - 1807
  • [50] Expanding variational autoencoders for learning and exploiting latent representations in search distributions
    Garciarena, Unai
    Santana, Roberto
    Mendiburu, Alexander
    GECCO'18: PROCEEDINGS OF THE 2018 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE, 2018, : 849 - 856