IENet: inheritance enhancement network for video salient object detection

被引:0
|
作者
Jiang, Tao [1 ]
Wang, Yi [2 ]
Hou, Feng [1 ]
Wang, Ruili [1 ]
机构
[1] Massey Univ, Sch Math & Computat Sci, Auckland 0632, New Zealand
[2] Dalian Univ Technol DUT, RU Int Sch Informat Sci & Engn, Dalian 116000, Peoples R China
基金
中国国家自然科学基金;
关键词
Video salient object detection; Feature fusion; Visual transformer; Frame-aware temporal relationships; OPTIMIZATION; CUES;
D O I
10.1007/s11042-024-18408-4
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Effective utilization of spatiotemporal information is essential for improving the accuracy and robustness of Video Salient Object Detection (V-SOD). However, current methods have not fully utilized historical frame information, ultimately resulting in insufficient integration of complementary semantic information. To address this issue, we propose a novel Inheritance Enhancement Network (IENet) based on Transformer. The core of IENet is a Heritable Multi-Frame Attention (HMA) module, which fully exploits long-term context and frame-aware temporal modeling in feature extraction through unidirectional cross-frame enhancement. In contrast to existing methods, our heritable strategy is based on the unidirectional inheritance model using attention maps which ensure the information propagation for each frame is consistent and orderly, avoiding additional interference. Furthermore, we propose an auxiliary attention loss by using inherited attention maps to direct the network to focus more on target regions. The experimental results of our IENet reveal its effectiveness in handling challenging scenes on five popular benchmark datasets. For instance, in the cases of VOS and DAVSOD, our method achieves 0.042% and 0.070% for MAE compared to other competitive models. Particularly, IENet excels in inheriting finer details from historical frames even in complex environments. The module and predicted maps are publicly available at https://github.com/TOMMYWHY/IENet
引用
收藏
页码:72007 / 72026
页数:20
相关论文
共 50 条
  • [41] Shifting More Attention to Video Salient Object Detection
    Fan, Deng-Ping
    Wang, Wenguan
    Cheng, Ming-Ming
    Shen, Jianbing
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 8546 - 8556
  • [42] Motion Guided Attention for Video Salient Object Detection
    Li, Haofeng
    Chen, Guanqi
    Li, Guanbin
    Yu, Yizhou
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 7273 - 7282
  • [43] Multiscale Feature Enhancement Network for Salient Object Detection in Optical Remote Sensing Images
    Wang, Zhen
    Guo, Jianxin
    Zhang, Chuanlei
    Wang, Buhong
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
  • [44] Depth cue enhancement and guidance network for RGB-D salient object detection
    Li, Xiang
    Zhang, Qing
    Yan, Weiqi
    Dai, Meng
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2023, 95
  • [45] Light field salient object detection network based on feature enhancement and mutual attention
    Zhu, Xi
    Xia, Huai
    Wang, Xucheng
    Zheng, Zhenrong
    JOURNAL OF ELECTRONIC IMAGING, 2024, 33 (05)
  • [46] Salient Object Detection With Spatiotemporal Background Priors for Video
    Xi, Tao
    Zhao, Wei
    Wang, Han
    Lin, Weisi
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2017, 26 (07) : 3425 - 3436
  • [47] Dual pyramid network for salient object detection
    Xu, Xuemiao
    Chen, Jiaxing
    Zhang, Huaidong
    Han, Guoqiang
    NEUROCOMPUTING, 2020, 375 : 113 - 123
  • [48] Decomposition and Completion Network for Salient Object Detection
    Wu, Zhe
    Su, Li
    Huang, Qingming
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 6226 - 6239
  • [49] Edge Enhancing Network for Salient Object Detection
    Zhao W.
    Wang H.
    Liu X.
    Tongji Daxue Xuebao/Journal of Tongji University, 2024, 52 (02): : 293 - 302
  • [50] Feature Recalibration Network for Salient Object Detection
    Tan, Zhenshan
    Gu, Xiaodong
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2022, PT IV, 2022, 13532 : 64 - 75