DAST: Depth-Aware Assessment and Synthesis Transformer for RGB-D Salient Object Detection

被引:1
|
作者
Xia, Chenxing [1 ]
Duan, Songsong [1 ]
Fang, Xianjin [1 ]
Ge, Bin [1 ]
Gao, Xiuju [1 ]
Cui, Jianhua [2 ]
机构
[1] Anhui Univ Sci & Technol, Huainan, Anhui, Peoples R China
[2] China Tobacco Henan Ind Co Ltd, Anyang Cigarette Factory, Anyang, Henan, Peoples R China
基金
美国国家科学基金会; 安徽省自然科学基金;
关键词
Salient object detection; Swin transformer; Low-quality; Depth map; Assessment and synthesis; NETWORK;
D O I
10.1007/978-3-031-20865-2_35
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The introduction and popularity of depth maps have brought new vitality and growth into salient object detection (SOD), and plentiful RGB-D SOD methods have been proposed, which mainly focus on how to utilize and integrate the depth map. Although existing methods have achieved promising performance, the negative effects of lowquality depth maps have not been effectively addressed. In this paper, we solve the problem with a strategy of judging low-quality depth maps and assigning low factors to low-quality depth maps. To this end, we proposed a novel Transformer-based SOD framework, namely Depth-aware Assessment and Synthesis Transformer (DAST), to further improve the performance of RGB-D SOD. The proposed DAST involves two primary designs: 1) a Swin Transformer-based encoder is employed instead of a convolutional neural network for more effective feature extraction and long-range dependencies capture; 2) a Depth Assessment and Synthesis (DAS) module is proposed to judge the quality of depth maps and fuse the multi-modality salient features by computing the difference of saliency maps from RGB and depth streams in a coarse-to-fine manner. Extensive experiments on five benchmark datasets demonstrate that the proposed DAST achieves favorable performance as compared with other state-of-the-art (SOTA) methods.
引用
收藏
页码:473 / 487
页数:15
相关论文
共 50 条
  • [1] Depth-aware lightweight network for RGB-D salient object detection
    Ling, Liuyi
    Wang, Yiwen
    Wang, Chengjun
    Xu, Shanyong
    Huang, Yourui
    [J]. IET IMAGE PROCESSING, 2023, 17 (08) : 2350 - 2361
  • [2] Depth-aware inverted refinement network for RGB-D salient object detection
    Gao, Lina
    Liu, Bing
    Fu, Ping
    Xu, Mingzhu
    [J]. NEUROCOMPUTING, 2023, 518 : 507 - 522
  • [3] OBJECT-AWARE CALIBRATED DEPTH-GUIDED TRANSFORMER FOR RGB-D CO-SALIENT OBJECT DETECTION
    Wu, Yang
    Liang, Lingyan
    Zhao, Yaqian
    Zhang, Kaihua
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 1121 - 1126
  • [4] Depth-Aware CNN for RGB-D Segmentation
    Wang, Weiyue
    Neumann, Ulrich
    [J]. COMPUTER VISION - ECCV 2018, PT XI, 2018, 11215 : 144 - 161
  • [5] SiaTrans: Siamese transformer network for RGB-D salient object detection with depth image classification
    Jia, XingZhao
    DongYe, ChangLei
    Peng, YanJun
    [J]. IMAGE AND VISION COMPUTING, 2022, 127
  • [6] GroupTransNet: Group transformer network for RGB-D salient object detection
    Fang, Xian
    Jiang, Mingfeng
    Zhu, Jinchao
    Shao, Xiuli
    Wang, Hongpeng
    [J]. NEUROCOMPUTING, 2024, 594
  • [7] MULTI-MODAL TRANSFORMER FOR RGB-D SALIENT OBJECT DETECTION
    Song, Peipei
    Zhang, Jing
    Koniusz, Piotr
    Barnes, Nick
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 2466 - 2470
  • [8] Context-aware network for RGB-D salient object detection
    Liang, Fangfang
    Duan, Lijuan
    Ma, Wei
    Qiao, Yuanhua
    Miao, Jun
    Ye, Qixiang
    [J]. PATTERN RECOGNITION, 2021, 111
  • [9] SwinNet: Swin Transformer Drives Edge-Aware RGB-D and RGB-T Salient Object Detection
    Liu, Zhengyi
    Tan, Yacheng
    He, Qian
    Xiao, Yun
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (07) : 4486 - 4497
  • [10] EM-Trans: Edge-Aware Multimodal Transformer for RGB-D Salient Object Detection
    Chen, Geng
    Wang, Qingyue
    Dong, Bo
    Ma, Ruitao
    Liu, Nian
    Fu, Huazhu
    Xia, Yong
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, : 1 - 14