Inversed Pyramid Network with Spatial-adapted and Task-oriented Tuning for few-shot learning

被引:0
|
作者
Zhao, Xiaowei [1 ]
Wang, Duorui [1 ]
Bai, Shihao [4 ]
Wang, Shuo [5 ]
Gao, Yajun [1 ]
Liang, Yu [6 ]
Ma, Yuqing [1 ,2 ]
Liu, Xianglong [1 ,3 ]
机构
[1] Beihang Univ, State Key Lab Complex & Crit Software Environm, Beijing 100191, Peoples R China
[2] Beihang Univ, Inst Artificial Intelligence, Beijing 100191, Peoples R China
[3] Zhongguancun Lab, Beijing 100194, Peoples R China
[4] SenseTime, Beijing 100080, Peoples R China
[5] Meituan, Beijing 100102, Peoples R China
[6] Beijing Univ Technol, Beijing 100124, Peoples R China
关键词
Few-shot learning; Inverted Pyramid Network; Spatial-adapted Layer; Task-oriented Tuning;
D O I
10.1016/j.patcog.2025.111415
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the rapid development of artificial intelligence, deep neural networks have achieved great performance in many tasks. However, traditional deep learning methods require a large amount of training data, which may not be available in certain practical scenarios. In contrast, few-shot learning aims to learn a model that can be readily adapted to new unseen classes from only one or a few labeled examples. Despite this success, most existing methods rely on pre-trained feature extractor networks trained with global features, ignoring the discrimination of local features, and weak generalization capabilities limit their performance. To address the problem, according to the human's coarse-to-fine cognition paradigm, we propose an Inverted Pyramid Network with Spatial-adapted and Task-oriented Tuning (TIPN) for few-shot learning. Specifically, the proposed framework represents local features for categories that are difficult to distinguish by global features and recognizes objects from both global and local perspectives. Moreover, to ensure the calibration validity of the proposed model at the local stage, we introduce the Spatial-adapted Layer to preserve the discriminative global representation ability of the pre-trained backbone network. Meanwhile, as the representations extracted from the past categories are not applicable to the current new tasks, we further propose the Task-oriented Tuning strategy to adjust the parameters of the Batch Normalization layer in the pre-trained feature extractor network, to explicitly transfer knowledge from base classes to novel classes according to the support samples of each task. Extensive experiments conducted on multiple benchmark datasets demonstrate that our method can significantly outperform many state-of-the-art few-shot learning methods.
引用
收藏
页数:11
相关论文
共 50 条
  • [31] Robust Compare Network for Few-Shot Learning
    Yang, Yixin
    Li, Yang
    Zhang, Rui
    Wang, Jiabao
    Miao, Zhuang
    IEEE ACCESS, 2020, 8 : 137966 - 137974
  • [32] Diversity Transfer Network for Few-Shot Learning
    Chen, Mengting
    Fang, Yuxin
    Wang, Xinggang
    Luo, Heng
    Geng, Yifeng
    Zhang, Xinyu
    Huang, Chang
    Liu, Wenyu
    Wang, Bo
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 10559 - 10566
  • [33] Learning to Compare: Relation Network for Few-Shot Learning
    Sung, Flood
    Yang, Yongxin
    Zhang, Li
    Xiang, Tao
    Torr, Philip H. S.
    Hospedales, Timothy M.
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 1199 - 1208
  • [34] Mutual Correlation Network for few-shot learning
    Chen, Derong
    Chen, Feiyu
    Ouyang, Deqiang
    Shao, Jie
    NEURAL NETWORKS, 2024, 175
  • [35] A Difference Measuring Network for Few-Shot Learning
    Wang, Yu
    Bao, Junpeng
    Li, Yanhua
    Feng, Zhonghui
    ARTIFICIAL INTELLIGENCE APPLICATIONS AND INNOVATIONS, AIAI 2023, PT II, 2023, 676 : 235 - 249
  • [36] Task Encoding With Distribution Calibration for Few-Shot Learning
    Zhang, Jing
    Zhang, Xinzhou
    Wang, Zhe
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (09) : 6240 - 6252
  • [37] Attentive matching network for few-shot learning
    Mai, Sijie
    Hu, Haifeng
    Xu, Jia
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2019, 187
  • [38] Robust Compare Network for Few-Shot Learning
    Yang, Yixin
    Li, Yang
    Zhang, Rui
    Wang, Jiabao
    Miao, Zhuang
    Zhang, Rui (3959966@qq.com), 1600, Institute of Electrical and Electronics Engineers Inc. (08): : 137966 - 137974
  • [39] Learning speed equalization network for few-shot learning
    Wang, Cailing
    Zhong, Chen
    Jiang, GuoPing
    JOURNAL OF ELECTRONIC IMAGING, 2022, 31 (01)
  • [40] Attention Relational Network for Few-Shot Learning
    Shuai, Jia
    Chen, JiaMing
    Yang, Meng
    INTELLIGENCE SCIENCE AND BIG DATA ENGINEERING: BIG DATA AND MACHINE LEARNING, PT II, 2019, 11936 : 163 - 174