Deep relational self-Attention networks for scene graph generation

被引:3
|
作者
Li, Ping [1 ]
Yu, Zhou [1 ]
Zhan, Yibing [1 ,2 ]
机构
[1] Hangzhou Dianzi Univ, Sch Comp Sci & Technol, Key Lab Complex Syst Modeling & Simulat, Hangzhou, Peoples R China
[2] JD Explore Acad, Beijing, Peoples R China
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
Scene graph generation; Image understanding; Deep neural networks;
D O I
10.1016/j.patrec.2021.12.013
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Scene graph generation (SGG) aims to simultaneously detect objects in an image and predict relations for these detected objects. SGG is challenging that requires modeling the contextualized relationships among objects rather than only considering relationships between paired objects. Most existing approaches ad -dress this problem by using a CNN or RNN framework, which can not explicitly and effectively models the dense interactions among objects. In this paper, we exploit the attention mechanism and introduce a relational self-attention (RSA) module to simultaneously model the object and relation contexts. By stack -ing such RSA modules in depth, we obtain a deep relational self-attention network (RSAN), which is able to characterize complex interactions thus facilitating the understanding of object and relation semantics. Extensive experiments on the benchmark Visual Genome dataset demonstrate the effectiveness of RSAN. (c) 2021 Elsevier B.V. All rights reserved.
引用
下载
收藏
页码:200 / 206
页数:7
相关论文
共 50 条
  • [21] Relational Self-Attention: What's Missing in Attention for Video Understanding
    Kim, Manjin
    Kwon, Heeseung
    Wang, Chunyu
    Kwak, Suha
    Cho, Minsu
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [22] Self-Attention Generative Adversarial Networks
    Zhang, Han
    Goodfellow, Ian
    Metaxas, Dimitris
    Odena, Augustus
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [23] Self-Attention Networks for Code Search
    Fang, Sen
    Tan, You-Shuai
    Zhang, Tao
    Liu, Yepang
    INFORMATION AND SOFTWARE TECHNOLOGY, 2021, 134
  • [24] Modeling Localness for Self-Attention Networks
    Yang, Baosong
    Tu, Zhaopeng
    Wong, Derek F.
    Meng, Fandong
    Chao, Lidia S.
    Zhang, Tong
    2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 4449 - 4458
  • [25] SAGSleepNet: A deep learning model for sleep staging based on self-attention graph of polysomnography
    Jin, Zheng
    Jia, Kebin
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2023, 86
  • [26] Very Deep Self-Attention Networks for End-to-End Speech Recognition
    Ngoc-Quan Pham
    Thai-Son Nguyen
    Niehues, Jan
    Mueller, Markus
    Waibel, Alex
    INTERSPEECH 2019, 2019, : 66 - 70
  • [27] Knowledge-Aware Self-Attention Networks for Document Grounded Dialogue Generation
    Tang, Xiangru
    Hu, Po
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, KSEM 2019, PT II, 2019, 11776 : 400 - 411
  • [28] Tailoring Self-Attention for Graph via Rooted Subtrees
    Huang, Siyuan
    Song, Yunchong
    Zhou, Jiayue
    Lin, Zhouhan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [29] A novel self-attention deep subspace clustering
    Chen, Zhengfan
    Ding, Shifei
    Hou, Haiwei
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2021, 12 (08) : 2377 - 2387
  • [30] Deep Semantic Role Labeling with Self-Attention
    Tan, Zhixing
    Wang, Mingxuan
    Xie, Jun
    Chen, Yidong
    Shi, Xiaodong
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 4929 - 4936