Visual identification of biological motion for underwater human-robot interaction

被引:2
|
作者
Sattar, Junaed [1 ]
Dudek, Gregory [2 ]
机构
[1] Univ Minnesota, Dept Comp Sci & Engn, Minneapolis, MN 55455 USA
[2] McGill Univ, Sch Comp Sci, Montreal, PQ, Canada
关键词
Underwater robotics; Visual servoing; Visual tracking; Diver tracking; Human-robot collaboration;
D O I
10.1007/s10514-017-9644-y
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present an algorithm for underwater robots to visually detect and track human motion. Our objective is to enable human-robot interaction by allowing a robot to follow behind a human moving in (up to) six degrees of freedom. In particular, we have developed a system to allow a robot to detect, track and follow a scuba diver by using frequency-domain detection of biological motion patterns. The motion of biological entities is characterized by combinations of periodic motions which are inherently distinctive. This is especially true of human swimmers. By using the frequency-space response of spatial signals over a number of video frames, we attempt to identify signatures pertaining to biological motion. This technique is applied to track scuba divers in underwater domains, typically with the robot swimming behind the diver. The algorithm is able to detect a range of motions, which includes motion directly away from or towards the camera. Once detected, the motion of the diver relative to the vehicle is then tracked using an Unscented Kalman Filter, an approach for non-linear estimation. The efficiency of our approach makes it attractive for real-time applications on-board our underwater vehicle, and in future applications we intend to track scuba divers in real-time with the robot. The paper presents an algorithmic overview of our approach, together with experimental evaluation based on underwater video footage.
引用
收藏
页码:111 / 124
页数:14
相关论文
共 50 条
  • [1] Visual identification of biological motion for underwater human–robot interaction
    Junaed Sattar
    Gregory Dudek
    [J]. Autonomous Robots, 2018, 42 : 111 - 124
  • [2] Robot Communication Via Motion: Closing the Underwater Human-Robot Interaction Loop
    Fulton, Michael
    Edge, Chelsey
    Sattar, Junaed
    [J]. 2019 INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2019, : 4660 - 4666
  • [3] Visual human-robot interaction
    Heinzmann, J
    Zelinsky, A
    [J]. 2001 INTERNATIONAL WORKSHOP ON BIO-ROBOTICS AND TELEOPERATION, PROCEEDINGS, 2001, : 113 - 118
  • [4] Narrative-level visual interpretation of human motion for human-robot interaction
    Hilti, A
    Nourbakhsh, I
    Jensen, B
    Siegwart, R
    [J]. IROS 2001: PROCEEDINGS OF THE 2001 IEEE/RJS INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, VOLS 1-4: EXPANDING THE SOCIETAL ROLE OF ROBOTICS IN THE NEXT MILLENNIUM, 2001, : 2074 - 2079
  • [5] An Underwater Human-Robot Interaction Using a Visual-Textual Model for Autonomous Underwater Vehicles
    Zhang, Yongji
    Jiang, Yu
    Qi, Hong
    Zhao, Minghao
    Wang, Yuehang
    Wang, Kai
    Wei, Fenglin
    [J]. SENSORS, 2023, 23 (01)
  • [6] Visual Perception for Multiple Human-Robot Interaction From Motion Behavior
    Benli, Emrah
    Motai, Yuichi
    Rogers, John
    [J]. IEEE SYSTEMS JOURNAL, 2020, 14 (02): : 2937 - 2948
  • [7] Advancements in Visual Gesture Recognition for Underwater Human-Robot Interaction: A Comprehensive Review
    Hozyn, Stanislaw
    [J]. IEEE Access, 2024, 12 : 163131 - 163142
  • [8] A Simulator for Underwater Human-Robot Interaction Scenarios
    DeMarco, Kevin J.
    West, Michael E.
    Howard, Ayanna M.
    [J]. 2013 OCEANS - SAN DIEGO, 2013,
  • [9] Visual Surveillance for Human-Robot Interaction
    Martinez-Martin, Ester
    del Pobil, Angel P.
    [J]. PROCEEDINGS 2012 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2012, : 3333 - 3338
  • [10] Understanding human motion and gestures for underwater human-robot collaboration
    Islam, Md Jahidul
    Ho, Marc
    Sattar, Junaed
    [J]. JOURNAL OF FIELD ROBOTICS, 2019, 36 (05) : 851 - 873