Visual identification of biological motion for underwater human–robot interaction

被引:0
|
作者
Junaed Sattar
Gregory Dudek
机构
[1] University of Minnesota,Department of Computer Science and Engineering
[2] McGill University,School of Computer Science
来源
Autonomous Robots | 2018年 / 42卷
关键词
Underwater robotics; Visual servoing; Visual tracking; Diver tracking; Human–robot collaboration;
D O I
暂无
中图分类号
学科分类号
摘要
We present an algorithm for underwater robots to visually detect and track human motion. Our objective is to enable human–robot interaction by allowing a robot to follow behind a human moving in (up to) six degrees of freedom. In particular, we have developed a system to allow a robot to detect, track and follow a scuba diver by using frequency-domain detection of biological motion patterns. The motion of biological entities is characterized by combinations of periodic motions which are inherently distinctive. This is especially true of human swimmers. By using the frequency-space response of spatial signals over a number of video frames, we attempt to identify signatures pertaining to biological motion. This technique is applied to track scuba divers in underwater domains, typically with the robot swimming behind the diver. The algorithm is able to detect a range of motions, which includes motion directly away from or towards the camera. Once detected, the motion of the diver relative to the vehicle is then tracked using an Unscented Kalman Filter, an approach for non-linear estimation. The efficiency of our approach makes it attractive for real-time applications on-board our underwater vehicle, and in future applications we intend to track scuba divers in real-time with the robot. The paper presents an algorithmic overview of our approach, together with experimental evaluation based on underwater video footage.
引用
收藏
页码:111 / 124
页数:13
相关论文
共 50 条
  • [1] Visual identification of biological motion for underwater human-robot interaction
    Sattar, Junaed
    Dudek, Gregory
    [J]. AUTONOMOUS ROBOTS, 2018, 42 (01) : 111 - 124
  • [2] Robot Communication Via Motion: Closing the Underwater Human-Robot Interaction Loop
    Fulton, Michael
    Edge, Chelsey
    Sattar, Junaed
    [J]. 2019 INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2019, : 4660 - 4666
  • [3] An Underwater Human-Robot Interaction Using a Visual-Textual Model for Autonomous Underwater Vehicles
    Zhang, Yongji
    Jiang, Yu
    Qi, Hong
    Zhao, Minghao
    Wang, Yuehang
    Wang, Kai
    Wei, Fenglin
    [J]. SENSORS, 2023, 23 (01)
  • [4] Narrative-level visual interpretation of human motion for human-robot interaction
    Hilti, A
    Nourbakhsh, I
    Jensen, B
    Siegwart, R
    [J]. IROS 2001: PROCEEDINGS OF THE 2001 IEEE/RJS INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, VOLS 1-4: EXPANDING THE SOCIETAL ROLE OF ROBOTICS IN THE NEXT MILLENNIUM, 2001, : 2074 - 2079
  • [5] Advancements in Visual Gesture Recognition for Underwater Human-Robot Interaction: A Comprehensive Review
    Hozyn, Stanislaw
    [J]. IEEE Access, 2024, 12 : 163131 - 163142
  • [6] Visual motion estimation for an autonomous underwater reef monitoring robot
    Dunbabin, Matthew
    Usher, Kane
    Corke, Peter
    [J]. FIELD AND SERVICE ROBOTICS, 2006, 25 : 31 - +
  • [7] Visual Perception for Multiple Human-Robot Interaction From Motion Behavior
    Benli, Emrah
    Motai, Yuichi
    Rogers, John
    [J]. IEEE SYSTEMS JOURNAL, 2020, 14 (02): : 2937 - 2948
  • [8] Visual human-robot interaction
    Heinzmann, J
    Zelinsky, A
    [J]. 2001 INTERNATIONAL WORKSHOP ON BIO-ROBOTICS AND TELEOPERATION, PROCEEDINGS, 2001, : 113 - 118
  • [9] Understanding human motion and gestures for underwater human-robot collaboration
    Islam, Md Jahidul
    Ho, Marc
    Sattar, Junaed
    [J]. JOURNAL OF FIELD ROBOTICS, 2019, 36 (05) : 851 - 873
  • [10] A Simulator for Underwater Human-Robot Interaction Scenarios
    DeMarco, Kevin J.
    West, Michael E.
    Howard, Ayanna M.
    [J]. 2013 OCEANS - SAN DIEGO, 2013,