Boosting Few-shot visual recognition via saliency-guided complementary attention

被引:7
|
作者
Zhao, Linglan [1 ]
Liu, Ge [1 ]
Guo, Dashan [1 ]
Li, Wei [1 ]
Fang, Xiangzhong [1 ]
机构
[1] Shanghai Jiao Tong Univ, Dept Elect Engn, Shanghai, Peoples R China
基金
日本学术振兴会;
关键词
Few-shot learning; Meta-learning; Spatial attention; Image classification; Saliency detection; NETWORK;
D O I
10.1016/j.neucom.2022.08.028
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Despite significant progress in recent deep neural networks, most deep learning algorithms rely heavily on abundant training samples. To address the issue, few-shot learning (FSL) methods are designed to learn models that can generalize to novel classes with limited training data. In this work, we propose an effective and interpretable FSL approach termed Saliency-Guided Complementary Attention (SGCA). Concretely, SGCA aims to boost few-shot visual recognition from two perspectives: learning generalizable feature representations and building a robust classification module in a unified framework. For general-izable representation learning, we propose to explore the intrinsic structure of natural images by training the feature extractor with an auxiliary task to segment foreground regions from background clutter. The guidance signals are provided during training by a saliency detector which highlights object regions in images corresponding to the human visual system. Moreover, for robust classification module building, we introduce a complementary attention mechanism based on the learned segmentation to make the classification module focus on various informative parts of the image. Extensive experiments on 5 pop-ular FSL datasets demonstrate that SGCA can outperform state-of-the-art approaches by a significant margin. In addition, extensions of SGCA to other challenging scenarios, including generalized, transduc-tive and semi-supervised FSL, also verify the effectiveness and flexibility of our proposed approach.(c) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页码:412 / 427
页数:16
相关论文
共 50 条
  • [1] Few-shot Learning via Saliency-guided Hallucination of Samples
    Zhang, Hongguang
    Zhang, Jing
    Koniusz, Piotr
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 2765 - 2774
  • [2] Saliency-guided meta-hallucinator for few-shot learning
    Zhang, Hongguang
    Liu, Chun
    Wang, Jiandong
    Ma, Linru
    Koniusz, Piotr
    Torr, Philip H. S.
    Yang, Lin
    SCIENCE CHINA-INFORMATION SCIENCES, 2024, 67 (10)
  • [3] Saliency-guided meta-hallucinator for few-shot learning
    Hongguang ZHANG
    Chun LIU
    Jiandong WANG
    Linru MA
    Piotr KONIUSZ
    Philip HSTORR
    Lin YANG
    Science China(Information Sciences), 2024, 67 (10) : 189 - 210
  • [4] Attributes-Guided and Pure-Visual Attention Alignment for Few-Shot Recognition
    Huang, Siteng
    Zhang, Min
    Kang, Yachen
    Wang, Donglin
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 7840 - 7847
  • [5] Boosting Factorization Machines via Saliency-Guided Mixup
    Wu, Chenwang
    Lian, Defu
    Ge, Yong
    Zhou, Min
    Chen, Enhong
    Tao, Dacheng
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (06) : 4443 - 4459
  • [6] SEGA: Semantic Guided Attention on Visual Prototype for Few-Shot Learning
    Yang, Fengyuan
    Wang, Ruiping
    Chen, Xilin
    2022 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2022), 2022, : 1586 - 1596
  • [7] VSA: Adaptive Visual and Semantic Guided Attention on Few-Shot Learning
    Chai, Jin
    Chen, Yisheng
    Shen, Weinan
    Zhang, Tong
    Chen, C. L. Philip
    ARTIFICIAL INTELLIGENCE, CICAI 2022, PT I, 2022, 13604 : 280 - 292
  • [8] Boosting Few-shot Action Recognition with Graph-guided Hybrid Matching
    Xing, Jiazheng
    Wang, Mengmeng
    Ruan, Yudi
    Chen, Bofan
    Guo, Yaowei
    Mu, Boyu
    Dai, Guang
    Wang, Jingdong
    Liu, Yong
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 1740 - 1750
  • [9] Saliency-guided neural prosthesis for visual attention: Design and simulation
    Yoshida, Masatoshi
    Veale, Richard
    NEUROSCIENCE RESEARCH, 2014, 78 : 90 - 94
  • [10] LEARNING SEMANTICS-GUIDED VISUAL ATTENTION FOR FEW-SHOT IMAGE CLASSIFICATION
    Chu, Wen-Hsuan
    Wang, Yu-Chiang Frank
    2018 25TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2018, : 2979 - 2983