Spatio-Temporal Action Order Representation for Mobile Manipulation Planning*

被引:0
|
作者
Kawasaki, Yosuke [1 ]
Takahashi, Masaki [2 ]
机构
[1] Keio Univ, Grad Sch Sci & Technol, 3-14-1 Hiyoshi,Kohoku Ku, Yokohama, Kanagawa 2238522, Japan
[2] Keio Univ, Dept Syst Design Engn, 3-14-1 Hiyoshi,Kohoku Ku, Yokohama, Kanagawa 2238522, Japan
基金
日本科学技术振兴机构;
关键词
TASK;
D O I
10.1109/RO-MAN53752.2022.9900643
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Social robots are used to perform mobile manipulation tasks, such as tidying up and carrying, based on instructions provided by humans. A mobile manipulation planner, which is used to exploit the robot's functions, requires a better understanding of the feasible actions in real space based on the robot's subsystem configuration and the object placement in the environment. This study aims to realize a mobile manipulation planner considering the world state, which consists of the robot state (subsystem configuration and their state) required to exploit the robot's functions. In this paper, this study proposes a novel environmental representation called a world state-dependent action graph (WDAG). The WDAG represents the spatial and temporal order of feasible actions based on the world state by adopting the knowledge representation with scene graphs and a recursive multilayered graph structure. The study also proposes a mobile manipulation planning method using the WDAG. The planner enables the derivation of many effective action sequences to accomplish the given tasks based on an exhaustive understanding of the spatial and temporal connections of actions. The effectiveness of the proposed method is evaluated through practical machine experiments performed. The experimental result demonstrates that the proposed method facilitates the effective utilization of the robot's functions.
引用
收藏
页码:1093 / 1098
页数:6
相关论文
共 50 条
  • [31] Action Tubelet Detector for Spatio-Temporal Action Localization
    Kalogeiton, Vicky
    Weinzaepfel, Philippe
    Ferrari, Vittorio
    Schmid, Cordelia
    2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 4415 - 4423
  • [32] Action Recognition by Learning Deep Multi-Granular Spatio-Temporal Video Representation
    Li, Qing
    Qiu, Zhaofan
    Yao, Ting
    Mei, Tao
    Rui, Yong
    Luo, Jiebo
    ICMR'16: PROCEEDINGS OF THE 2016 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, 2016, : 159 - 166
  • [33] Spatio-temporal feature extraction and representation for RGB-D human action recognition
    Luo, Jiajia
    Wang, Wei
    Qi, Hairong
    PATTERN RECOGNITION LETTERS, 2014, 50 : 139 - 148
  • [34] Probabilistic Spatio-Temporal Fusion of Affordances for Grasping and Manipulation
    Pohl, Christoph
    Asfour, Tamim
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (02) : 3226 - 3233
  • [35] Action matching network: open-set action recognition using spatio-temporal representation matching
    Yu, Jongmin
    Kim, Du Yong
    Yoon, Yongsang
    Jeon, Moongu
    VISUAL COMPUTER, 2020, 36 (07): : 1457 - 1471
  • [36] Action matching network: open-set action recognition using spatio-temporal representation matching
    Jongmin Yu
    Du Yong Kim
    Yongsang Yoon
    Moongu Jeon
    The Visual Computer, 2020, 36 : 1457 - 1471
  • [37] Spatio-Temporal Coordination of Mobile Robot Swarms
    Graff, Daniel
    Karnapke, Reinhardt
    2016 IEEE 41ST CONFERENCE ON LOCAL COMPUTER NETWORKS (LCN), 2016, : 511 - 514
  • [38] Querying mobile objects in spatio-temporal databases
    Porkaew, K
    Lazaridis, I
    Mehrotra, S
    ADVANCES IN SPATIAL AND TEMPORAL DATABASES, PROCEEDINGS, 2001, 2121 : 59 - 78
  • [39] Collaborative assistance with spatio-temporal planning problems
    Seifert, Inessa
    SPATIAL COGNITION V: REASONING, ACTION, INTERACTION, 2007, 4387 : 90 - 106
  • [40] Planning with Spatio-Temporal Search Control Knowledge
    Lu, Xu
    Tian, Cong
    Duan, Zhenhua
    Du, Hongwei
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2018, 30 (10) : 1915 - 1928