PileNet: A high-and-low pass complementary filter with multi-level feature refinement for salient object detection

被引:0
|
作者
Yang, Xiaoqi [1 ]
Duan, Liangliang [1 ]
Zhou, Quanqiang [1 ]
机构
[1] Qingdao Univ Technol, Sch Informat & Control Engn, Qingdao 266520, Peoples R China
关键词
Salient object detection; Complementary filter; Feature refinement; Feature aggregation; NETWORK;
D O I
10.1016/j.jvcir.2024.104186
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Multi-head self-attentions (MSAs) in Transformer are low-pass filters, which will tend to reduce high-frequency signals. Convolutional layers (Convs) in Convolutional Neural Network (CNN) are high-pass filters, which will tend to capture high-frequency components of the images. Therefore, CNN and Transformer contain complementary information, and the combination of the two is necessary for satisfactory detection results. In this work, we propose a novel framework PileNet that efficiently combine CNN and Transformer for accurate salient object detection (SOD). Specifically in PileNet, we introduce complementary encoder that extracts multi-level complementary saliency features. Next, we simplify the complementary features by adjusting the number of channels for all features to a fixed value. By introducing the multi-level feature aggregation (MLFA) and multi-level feature refinement (MLFR) units, the lowand high-level features can easily be transmitted to feature blocks at various pyramid levels. Finally, we fuse all the refined saliency features in a Unet-like structure from top to bottom and use multi-point supervision mechanism to produce the final saliency maps. Extensive experimental results over five widely used saliency benchmark datasets clearly demonstrate that our proposed model can accurately locate the entire salient objects with clear object boundaries and outperform sixteen previous state-of-the-art saliency methods in terms of a wide range of metrics.
引用
收藏
页数:10
相关论文
共 50 条
  • [31] EMNet: Edge-guided multi-level network for salient object detection in low-light images
    Jing, Lianghu
    Wang, Bo
    [J]. IMAGE AND VISION COMPUTING, 2024, 143
  • [32] Multi-level feature enhancement network for object detection in sonar images
    Zhou, Xin
    Zhou, Zihan
    Wang, Manying
    Ning, Bo
    Wang, Yanhao
    Zhu, Pengli
    [J]. JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2024, 100
  • [33] Bmsmlet: boosting multi-scale information on multi-level aggregated features for salient object detection
    Wu, Ziwei
    Jia, Tong
    Wu, Yunhe
    Zeng, Zhikang
    Liang, Feng
    [J]. VISUAL COMPUTER, 2024, 40 (02): : 1131 - 1144
  • [34] Bmsmlet: boosting multi-scale information on multi-level aggregated features for salient object detection
    Ziwei Wu
    Tong Jia
    Yunhe Wu
    Zhikang Zeng
    Feng Liang
    [J]. The Visual Computer, 2024, 40 (2) : 1131 - 1144
  • [35] Multi-level feature fusion and joint refinement for simultaneous object pose estimation and camera localization
    Wang, Junyi
    Qi, Yue
    [J]. NEURAL NETWORKS, 2024, 174
  • [36] Attention guided multi-level feature aggregation network for camouflaged object detection
    Wang, Anzhi
    Ren, Chunhong
    Zhao, Shuang
    Mu, Shibiao
    [J]. IMAGE AND VISION COMPUTING, 2024, 144
  • [37] Multi-Level Context Aggregation Network with Channel-Wise Attention for Salient Object Detection
    Jia, Zihui
    Weng, Zhenyu
    Wan, Fang
    Zhu, Yuesheng
    [J]. IEEE Access, 2020, 8 : 102303 - 102312
  • [38] RGB-T Salient Object Detection via Fusing Multi-Level CNN Features
    Zhang, Qiang
    Huang, Nianchang
    Yao, Lin
    Zhang, Dingwen
    Shan, Caifeng
    Han, Jungong
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 : 3321 - 3335
  • [39] Multi-Level Context Aggregation Network With Channel-Wise Attention for Salient Object Detection
    Jia, Zihui
    Weng, Zhenyu
    Wan, Fang
    Zhu, Yuesheng
    [J]. IEEE ACCESS, 2020, 8 : 102303 - 102312
  • [40] Evaluating salient object detection in natural images with multiple objects having multi-level saliency
    Yildirim, Goekhan
    Sen, Debashis
    Kankanhalli, Mohan
    Suesstrunk, Sabine
    [J]. IET IMAGE PROCESSING, 2020, 14 (10) : 2249 - 2262