SAN: Structure-aware attention network for dyadic human relation recognition in images

被引:0
|
作者
Kaen Kogashi
Shohei Nobuhara
Ko Nishino
机构
[1] Kyoto University,Department of Intelligence Science and Technology, Graduate School of Informatics
来源
关键词
Dyadic human relation recognition (DHR); DHR dataset; Multi-task learning;
D O I
暂无
中图分类号
学科分类号
摘要
We introduce a new dataset and method for Dyadic Human relation Recognition (DHR). DHR is a new task that concerns the recognition of the type (i.e., verb) and roles of a two-person interaction. Unlike past human action detection, our goal is to extract richer information regarding the roles of actors, i.e., which subjective person is acting on which objective person. For this, we introduce the DHR-WebImages dataset which consists of a total of 22,046 images of 51 verb classes of DHR with per-image annotation of the verb and role, and also a test set for evaluating generalization capabilities, which we refer to as DHR-Generalization. We tackle DHR by introducing a novel network inspired by the hierarchical nature of cognitive human perception. At the core of the network lies a “structure-aware attention” module that weights and integrates various hierarchical visual cues associated with the DHR instance in the image. The feature hierarchy consists of three levels, namely the union, human, and joint levels, each of which extracts visual features relevant to the participants while modeling their cross-talk. We refer to this network as Structure-aware Attention Network (SAN). Experimental results show that SAN achieves accurate DHR robust to lacking visibility of actors, and outperforms past methods by 3.04 mAP on DHR-WebImages verb task.
引用
下载
收藏
页码:46947 / 46966
页数:19
相关论文
共 50 条
  • [1] SAN: Structure-aware attention network for dyadic human relation recognition in images
    Kogashi, Kaen
    Nobuhara, Shohei
    Nishino, Ko
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 83 (16) : 46947 - 46966
  • [2] Relation Structure-Aware Heterogeneous Graph Neural Network
    Zhu, Shichao
    Zhou, Chuan
    Pan, Shirui
    Zhu, Xingquan
    Wang, Bin
    2019 19TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2019), 2019, : 1534 - 1539
  • [3] Relation Structure-Aware Heterogeneous Information Network Embedding
    Lu, Yuanfu
    Shi, Chuan
    Hu, Linmei
    Liu, Zhiyuan
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 4456 - 4463
  • [4] A Structure-Aware Relation Network for Thoracic Diseases Detection and Segmentation
    Lian, Jie
    Liu, Jingyu
    Zhang, Shu
    Gao, Kai
    Liu, Xiaoqing
    Zhang, Dingwen
    Yu, Yizhou
    IEEE TRANSACTIONS ON MEDICAL IMAGING, 2021, 40 (08) : 2042 - 2052
  • [5] RHINE: Relation Structure-Aware Heterogeneous Information Network Embedding
    Shi, Chuan
    Lu, Yuanfu
    Hu, Linmei
    Liu, Zhiyuan
    Ma, Huadong
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2022, 34 (01) : 433 - 447
  • [6] A Heterogeneous Graph Neural Network With Attribute Enhancement and Structure-Aware Attention
    Fan, Shenghang
    Liu, Guanjun
    Li, Jian
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2024, 11 (01): : 829 - 838
  • [7] Structure-Aware Image Expansion with Global Attention
    Guo, Dewen
    Feng, Jie
    Zhou, Bingfeng
    SA'19: SIGGRAPH ASIA 2019 TECHNICAL BRIEFS, 2019, : 13 - 16
  • [8] Enabling inductive knowledge graph completion via structure-aware attention network
    Wang, Jingchao
    Li, Weimin
    Liu, Wei
    Wang, Can
    Jin, Qun
    APPLIED INTELLIGENCE, 2023, 53 (21) : 25003 - 25027
  • [9] Enabling inductive knowledge graph completion via structure-aware attention network
    Jingchao Wang
    Weimin Li
    Wei Liu
    Can Wang
    Qun Jin
    Applied Intelligence, 2023, 53 : 25003 - 25027
  • [10] Structure-Aware Multiscale Hybrid Network for Change Detection of Remote Sensing Images
    Liu Qi
    Cao Lin
    Tian Shu
    Du Kangning
    Song Peiran
    Guo Yanan
    LASER & OPTOELECTRONICS PROGRESS, 2024, 61 (14)