TANet: Transformer-based asymmetric network for RGB-D salient object detection

被引:6
|
作者
Liu, Chang [1 ]
Yang, Gang [1 ,3 ]
Wang, Shuo [1 ]
Wang, Hangxu [1 ,2 ]
Zhang, Yunhua [1 ]
Wang, Yutao [1 ]
机构
[1] Northeastern Univ, Shenyang, Liaoning, Peoples R China
[2] DUT Artificial Intelligence Inst, Dalian, Peoples R China
[3] Northeastern Univ, Wenhua Rd, Shenyang 110000, Liaoning, Peoples R China
基金
中国国家自然科学基金;
关键词
computer vision; image segmentation; object detection; REGION;
D O I
10.1049/cvi2.12177
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Existing RGB-D salient object detection methods mainly rely on a symmetric two-stream Convolutional Neural Network (CNN)-based network to extract RGB and depth channel features separately. However, there are two problems with the symmetric conventional network structure: first, the ability of CNN in learning global contexts is limited; second, the symmetric two-stream structure ignores the inherent differences between modalities. In this study, a Transformer-based asymmetric network is proposed to tackle the issues mentioned above. The authors employ the powerful feature extraction capability of Transformer to extract global semantic information from RGB data and design a lightweight CNN backbone to extract spatial structure information from depth data without pre-training. The asymmetric hybrid encoder effectively reduces the number of parameters in the model while increasing speed without sacrificing performance. Then, a cross-modal feature fusion module which enhances and fuses RGB and depth features with each other is designed. Finally, the authors add edge prediction as an auxiliary task and propose an edge enhancement module to generate sharper contours. Extensive experiments demonstrate that our method achieves superior performance over 14 state-of-the-art RGB-D methods on six public datasets. The code of the authors will be released at .
引用
收藏
页码:415 / 430
页数:16
相关论文
共 50 条
  • [41] Salient object detection for RGB-D images by generative adversarial network
    Liu, Zhengyi
    Tang, Jiting
    Xiang, Qian
    Zhao, Peng
    MULTIMEDIA TOOLS AND APPLICATIONS, 2020, 79 (35-36) : 25403 - 25425
  • [42] An adaptive guidance fusion network for RGB-D salient object detection
    Haodong Sun
    Yu Wang
    Xinpeng Ma
    Signal, Image and Video Processing, 2024, 18 : 1683 - 1693
  • [43] Adaptive Depth Enhancement Network for RGB-D Salient Object Detection
    Yi, Kang
    Li, Yumeng
    Tang, Haoran
    Xu, Jing
    IEEE SIGNAL PROCESSING LETTERS, 2025, 32 : 176 - 180
  • [44] Salient object detection for RGB-D images by generative adversarial network
    Zhengyi Liu
    Jiting Tang
    Qian Xiang
    Peng Zhao
    Multimedia Tools and Applications, 2020, 79 : 25403 - 25425
  • [45] RGB-D salient object detection with asymmetric cross-modal fusion
    Yu M.
    Xing Z.-H.
    Liu Y.
    Kongzhi yu Juece/Control and Decision, 2023, 38 (09): : 2487 - 2495
  • [46] CSNet: a ConvNeXt-based Siamese network for RGB-D salient object detection
    Yunhua Zhang
    Hangxu Wang
    Gang Yang
    Jianhao Zhang
    Congjin Gong
    Yutao Wang
    The Visual Computer, 2024, 40 : 1805 - 1823
  • [47] A cascaded refined rgb-d salient object detection network based on the attention mechanism
    Zong, Guanyu
    Wei, Longsheng
    Guo, Siyuan
    Wang, Yongtao
    APPLIED INTELLIGENCE, 2023, 53 (11) : 13527 - 13548
  • [48] A cascaded refined rgb-d salient object detection network based on the attention mechanism
    Guanyu Zong
    Longsheng Wei
    Siyuan Guo
    Yongtao Wang
    Applied Intelligence, 2023, 53 : 13527 - 13548
  • [49] CSNet: a ConvNeXt-based Siamese network for RGB-D salient object detection
    Zhang, Yunhua
    Wang, Hangxu
    Yang, Gang
    Zhang, Jianhao
    Gong, Congjin
    Wang, Yutao
    VISUAL COMPUTER, 2024, 40 (03): : 1805 - 1823
  • [50] Transformer-Based Cross-Modal Integration Network for RGB-T Salient Object Detection
    Lv, Chengtao
    Zhou, Xiaofei
    Wan, Bin
    Wang, Shuai
    Sun, Yaoqi
    Zhang, Jiyong
    Yan, Chenggang
    IEEE TRANSACTIONS ON CONSUMER ELECTRONICS, 2024, 70 (02) : 4741 - 4755