Automating Gaze Target Annotation in Human-Robot Interaction

被引:0
|
作者
Cheng, Linlin [1 ]
Hindriks, Koen V. [1 ]
Belopolsky, Artem V. [2 ]
机构
[1] Vrije Univ Amsterdam, Fac Sci, Comp Sci, Amsterdam, Netherlands
[2] Vrije Univ Amsterdam, Dept Human Movement Sci, Amsterdam, Netherlands
关键词
D O I
10.1109/RO-MAN60168.2024.10731455
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Identifying gaze targets in videos of human-robot interaction is useful for measuring engagement. In practice, this requires manually annotating for a fixed set of objects that a participant is looking at in a video, which is very time-consuming. To address this issue, we propose an annotation pipeline for automating this effort. In this work, we focus on videos in which the objects looked at do not move. As input for the proposed pipeline, we therefore only need to annotate object bounding boxes for the first frame of each video. The benefit, moreover, of manually annotating these frames is that we can also draw bounding boxes for objects outside of it, which enables estimating gaze targets in videos where not all objects are visible. A second issue that we address is that the models used for automating the pipeline annotate individual video frames. In practice, however, manual annotation is done at the event level for video segments instead of single frames. Therefore, we also introduce and investigate several variants of algorithms for aggregating frame-level to event-level annotations, which are used in the last step in our annotation pipeline. We compare two versions of our pipeline: one that uses a state-of-the-art gaze estimation model (GEM) and a second one using a state-of-the-art target detection model (TDM). Our results show that both versions successfully automate the annotation, but the GEM pipeline performs slightly (approximate to 10%) better for videos where not all objects are visible. Analysis of our aggregation algorithm, moreover, shows that there is no need for manual video segmentation because a fixed time interval for segmentation yields very similar results. We conclude that the proposed pipeline can be used to automate almost all of the annotation effort.
引用
收藏
页码:991 / 998
页数:8
相关论文
共 50 条
  • [21] Human-Robot Interaction
    Jia, Yunyi
    Zhang, Biao
    Li, Miao
    King, Brady
    Meghdari, Ali
    JOURNAL OF ROBOTICS, 2018, 2018
  • [22] Human-Robot Interaction
    Sidobre, Daniel
    Broquere, Xavier
    Mainprice, Jim
    Burattini, Ernesto
    Finzi, Alberto
    Rossi, Silvia
    Staffa, Mariacarla
    ADVANCED BIMANUAL MANIPULATION: RESULTS FROM THE DEXMART PROJECT, 2012, 80 : 123 - +
  • [23] Human-Robot Interaction
    Sethumadhavan, Arathi
    ERGONOMICS IN DESIGN, 2012, 20 (03) : 27 - +
  • [24] Human-robot interaction
    Murphy R.R.
    Nomura T.
    Billard A.
    Burke J.L.
    IEEE Robotics and Automation Magazine, 2010, 17 (02): : 85 - 89
  • [25] Human-robot interaction
    Kosuge, K
    Hirata, Y
    IEEE ROBIO 2004: PROCEEDINGS OF THE IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS, 2004, : 8 - 11
  • [26] Human-robot interaction
    Sidobre, Daniel
    Broquère, Xavier
    Mainprice, Jim
    Burattini, Ernesto
    Finzi, Alberto
    Rossi, Silvia
    Staffa, Mariacarla
    Springer Tracts in Advanced Robotics, 2012, 80 (STAR): : 123 - 172
  • [27] Human-Robot Interaction
    Ivaldi, Serena
    Pateraki, Maria
    ERCIM NEWS, 2018, (114): : 6 - 7
  • [28] Gaze-Aware Social Interaction Techniques for Human-Robot Collaborative Shopping
    Iwasaki, Masaya
    Ogawa, Kosuke
    Kawamura, Tatsuyuki
    Nakanishi, Hideyuki
    COLLABORATION TECHNOLOGIES AND SOCIAL COMPUTING, COLLABTECH 2023, 2023, 14199 : 193 - 200
  • [29] Eye Gaze and Hand Gesture-Driven Human-Robot Interaction in Construction
    Wang, Xin
    Veeramani, Dharmaraj
    Dai, Fei
    Zhu, Zhenhua
    CONSTRUCTION RESEARCH CONGRESS 2024: ADVANCED TECHNOLOGIES, AUTOMATION, AND COMPUTER APPLICATIONS IN CONSTRUCTION, 2024, : 991 - 1000
  • [30] GazeEMD: Detecting Visual Intention in Gaze-Based Human-Robot Interaction
    Shi, Lei
    Copot, Cosmin
    Vanlanduit, Steve
    ROBOTICS, 2021, 10 (02)