Visual-Semantic Graph Attention Networks for Human-Object Interaction Detection

被引:13
|
作者
Liang, Zhijun [1 ]
Liu, Junfa [1 ]
Guan, Yisheng [1 ]
Rojas, Juan [2 ]
机构
[1] Guangdong Univ Technol, Sch Electromech Engn, Biomimet & Intelligent Robot Lab BIRL, Guangzhou 510006, Peoples R China
[2] Chinese Univ Hong Kong, Dept Mech & Automat Engn, Hong Kong, Peoples R China
关键词
D O I
10.1109/ROBIO54168.2021.9739429
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In scene understanding, robots benefit from not only detecting individual scene instances but also from learning their possible interactions. Human-Object Interaction (HOI) Detection infers the action predicate on a <human, predicate, object> triplet. Contextual information has been found critical in inferring interactions. However, most works only use local features from single human-object pairs for inference. Few works have studied the disambiguating contribution of subsidiary relations made available via graph networks. Similarly, few have leveraged visual cues with the intrinsic semantic regularities embedded in HOIs. We contribute Visual-Semantic Graph Attention Networks (VS-GATs): a dual-graph attention network that effectively aggregates visual, spatial, and semantic contextual information dynamically from primary human-object relations as well as subsidiary relations through attention mechanisms for strong disambiguating power. We achieve competitive results on two benchmarks: V-COCO and HICO-DET. The code is available at https://github.com/ birlrobotics/vs- gats.
引用
收藏
页码:1441 / 1447
页数:7
相关论文
共 50 条
  • [1] Human-object interaction detection via interactive visual-semantic graph learning
    Tongtong WU
    Fuqing DUAN
    Liang CHANG
    Ke LU
    Science China(Information Sciences), 2022, 65 (06) : 81 - 82
  • [2] Human-object interaction detection via interactive visual-semantic graph learning
    Wu, Tongtong
    Duan, Fuqing
    Chang, Liang
    Lu, Ke
    SCIENCE CHINA-INFORMATION SCIENCES, 2022, 65 (06)
  • [3] Human-object interaction detection via interactive visual-semantic graph learning
    Tongtong Wu
    Fuqing Duan
    Liang Chang
    Ke Lu
    Science China Information Sciences, 2022, 65
  • [4] SKGHOI: Spatial-Semantic Knowledge Graph for Human-Object Interaction Detection
    Zhu, Lijing
    Lan, Qizhen
    Velasquez, Alvaro
    Song, Houbing
    Kamal, Acharya
    Tian, Qing
    Niu, Shuteng
    2023 23RD IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS, ICDMW 2023, 2023, : 1186 - 1193
  • [5] Pose attention and object semantic representation-based human-object interaction detection network
    Deng, Wei-Mo
    Zhang, Hong-Bo
    Lei, Qing
    Du, Ji-Xiang
    Huang, Min
    MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 81 (27) : 39453 - 39470
  • [6] Pose attention and object semantic representation-based human-object interaction detection network
    Wei-Mo Deng
    Hong-Bo Zhang
    Qing Lei
    Ji-Xiang Du
    Min Huang
    Multimedia Tools and Applications, 2022, 81 : 39453 - 39470
  • [7] Semantic Inference Network for Human-Object Interaction Detection
    Liu, Hongyi
    Mo, Lisha
    Ma, Huimin
    IMAGE AND GRAPHICS, ICIG 2019, PT I, 2019, 11901 : 518 - 529
  • [8] Human-Object Interaction Detection Based on Star Graph
    Cai, Shuang
    Ma, Shiwei
    Gu, Dongzhou
    Wang, Chang
    INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2022, 36 (09)
  • [9] Deep Contextual Attention for Human-Object Interaction Detection
    Wang, Tiancai
    Anwer, Rao Muhammad
    Khan, Muhammad Haris
    Khan, Fahad Shahbaz
    Pang, Yanwei
    Shao, Ling
    Laaksonen, Jorma
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 5693 - 5701
  • [10] Pose graph parsing network for human-object interaction detection
    Su, Zhan
    Wang, Yuting
    Xie, Qing
    Yu, Ruiyun
    NEUROCOMPUTING, 2022, 476 : 53 - 62