Deep relational self-Attention networks for scene graph generation

被引:3
|
作者
Li, Ping [1 ]
Yu, Zhou [1 ]
Zhan, Yibing [1 ,2 ]
机构
[1] Hangzhou Dianzi Univ, Sch Comp Sci & Technol, Key Lab Complex Syst Modeling & Simulat, Hangzhou, Peoples R China
[2] JD Explore Acad, Beijing, Peoples R China
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
Scene graph generation; Image understanding; Deep neural networks;
D O I
10.1016/j.patrec.2021.12.013
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Scene graph generation (SGG) aims to simultaneously detect objects in an image and predict relations for these detected objects. SGG is challenging that requires modeling the contextualized relationships among objects rather than only considering relationships between paired objects. Most existing approaches ad -dress this problem by using a CNN or RNN framework, which can not explicitly and effectively models the dense interactions among objects. In this paper, we exploit the attention mechanism and introduce a relational self-attention (RSA) module to simultaneously model the object and relation contexts. By stack -ing such RSA modules in depth, we obtain a deep relational self-attention network (RSAN), which is able to characterize complex interactions thus facilitating the understanding of object and relation semantics. Extensive experiments on the benchmark Visual Genome dataset demonstrate the effectiveness of RSAN. (c) 2021 Elsevier B.V. All rights reserved.
引用
下载
收藏
页码:200 / 206
页数:7
相关论文
共 50 条
  • [31] Graph neural network with self-attention for material discovery
    Chen, Xuesi
    Jiang, Hantong
    Lin, Xuanjie
    Ren, Yongsheng
    Wu, Congzhong
    Zhan, Shu
    Ma, Wenhui
    MOLECULAR PHYSICS, 2023, 121 (04)
  • [32] SELF-ATTENTION EQUIPPED GRAPH CONVOLUTIONS FOR DISEASE PREDICTION
    Kazi, Anees
    Krishna, S. Arvind
    Shekarforoush, Shayan
    Kortuem, Karsten
    Albarqouni, Shadi
    Navab, Nassir
    2019 IEEE 16TH INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING (ISBI 2019), 2019, : 1896 - 1899
  • [33] Keyphrase Generation Based on Self-Attention Mechanism
    Yang, Kehua
    Wang, Yaodong
    Zhang, Wei
    Yao, Jiqing
    Le, Yuquan
    CMC-COMPUTERS MATERIALS & CONTINUA, 2019, 61 (02): : 569 - 581
  • [34] Self-Attention Mechanism in GANs for Molecule Generation
    Chinnareddy, Sandeep
    Grandhi, Pranav
    Narayan, Apurva
    20TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA 2021), 2021, : 57 - 60
  • [35] The Multimodal Scene Recognition Method Based on Self-Attention and Distillation
    Sun N.
    Xu W.
    Liu J.
    Chai L.
    Sun H.
    IEEE Multimedia, 2024, 31 (04) : 1 - 11
  • [36] Deep CNNs With Self-Attention for Speaker Identification
    Nguyen Nang An
    Nguyen Quang Thanh
    Liu, Yanbing
    IEEE ACCESS, 2019, 7 : 85327 - 85337
  • [37] A novel self-attention deep subspace clustering
    Zhengfan Chen
    Shifei Ding
    Haiwei Hou
    International Journal of Machine Learning and Cybernetics, 2021, 12 : 2377 - 2387
  • [38] Compressed Self-Attention for Deep Metric Learning
    Chen, Ziye
    Gong, Mingming
    Xu, Yanwu
    Wang, Chaohui
    Zhang, Kun
    Du, Bo
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 3561 - 3568
  • [39] HIGSA: Human image generation with self-attention
    Wu, Haoran
    He, Fazhi
    Si, Tongzhen
    Duan, Yansong
    Yan, Xiaohu
    ADVANCED ENGINEERING INFORMATICS, 2023, 55
  • [40] LayoutTransformer: Layout Generation and Completion with Self-attention
    Gupta, Kamal
    Lazarow, Justin
    Achille, Alessandro
    Davis, Larry
    Mahadevan, Vijay
    Shrivastava, Abhinav
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 984 - 994