Deep relational self-Attention networks for scene graph generation

被引:3
|
作者
Li, Ping [1 ]
Yu, Zhou [1 ]
Zhan, Yibing [1 ,2 ]
机构
[1] Hangzhou Dianzi Univ, Sch Comp Sci & Technol, Key Lab Complex Syst Modeling & Simulat, Hangzhou, Peoples R China
[2] JD Explore Acad, Beijing, Peoples R China
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
Scene graph generation; Image understanding; Deep neural networks;
D O I
10.1016/j.patrec.2021.12.013
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Scene graph generation (SGG) aims to simultaneously detect objects in an image and predict relations for these detected objects. SGG is challenging that requires modeling the contextualized relationships among objects rather than only considering relationships between paired objects. Most existing approaches ad -dress this problem by using a CNN or RNN framework, which can not explicitly and effectively models the dense interactions among objects. In this paper, we exploit the attention mechanism and introduce a relational self-attention (RSA) module to simultaneously model the object and relation contexts. By stack -ing such RSA modules in depth, we obtain a deep relational self-attention network (RSAN), which is able to characterize complex interactions thus facilitating the understanding of object and relation semantics. Extensive experiments on the benchmark Visual Genome dataset demonstrate the effectiveness of RSAN. (c) 2021 Elsevier B.V. All rights reserved.
引用
下载
收藏
页码:200 / 206
页数:7
相关论文
共 50 条
  • [41] Context-Aware Self-Attention Networks
    Yang, Baosong
    Li, Jian
    Wong, Derek F.
    Chao, Lidia S.
    Wang, Xing
    Tu, Zhaopeng
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 387 - 394
  • [42] Local Structural Aware Heterogeneous Information Network Embedding Based on Relational Self-Attention Graph Neural Network
    Cao, Meng
    Yuan, Jinliang
    Xu, Ming
    Yu, Hualei
    Wang, Chongjun
    IEEE ACCESS, 2021, 9 : 88301 - 88312
  • [43] Feature Importance Estimation with Self-Attention Networks
    Skrlj, Blaz
    Dzeroski, Saso
    Lavrac, Nada
    Petkovic, Matej
    ECAI 2020: 24TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, 325 : 1491 - 1498
  • [44] Improving Self-Attention Networks With Sequential Relations
    Zheng, Zaixiang
    Huang, Shujian
    Weng, Rongxiang
    Dai, Xinyu
    Chen, Jiajun
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2020, 28 : 1707 - 1716
  • [45] Multi-view self-attention networks
    Xu, Mingzhou
    Yang, Baosong
    Wong, Derek F.
    Chao, Lidia S.
    KNOWLEDGE-BASED SYSTEMS, 2022, 241
  • [46] Leveraging multimodal features for knowledge graph entity alignment based on dynamic self-attention networks
    Qian, Ye
    Pan, Li
    EXPERT SYSTEMS WITH APPLICATIONS, 2023, 228
  • [47] Semi-supervised Training for Knowledge Base Graph Self-attention Networks on Link Prediction
    Yao, Shuanglong
    Pi, Dechang
    Chen, Junfu
    Liu, Yufei
    Wu, Zhiyuan
    arXiv, 2022,
  • [48] SA-SinGAN: self-attention for single-image generation adversarial networks
    Chen, Xi
    Zhao, Hongdong
    Yang, Dongxu
    Li, Yueyuan
    Kang, Qing
    Lu, Haiyan
    MACHINE VISION AND APPLICATIONS, 2021, 32 (04)
  • [49] Multi-relational graph attention networks for knowledge graph completion
    Li, Zhifei
    Zhao, Yue
    Zhang, Yan
    Zhang, Zhaoli
    KNOWLEDGE-BASED SYSTEMS, 2022, 251
  • [50] Simple and deep graph attention networks
    Su, Guangxin
    Wang, Hanchen
    Zhang, Ying
    Zhang, Wenjie
    Lin, Xuemin
    KNOWLEDGE-BASED SYSTEMS, 2024, 293