Few-Shot Learning of Compact Models via Task-Specific Meta Distillation

被引:2
|
作者
Wu, Yong [1 ]
Chanda, Shekhor [2 ]
Hosseinzadeh, Mehrdad [3 ]
Liu, Zhi [1 ]
Wang, Yang [4 ]
机构
[1] Shanghai Univ, Shanghai, Peoples R China
[2] Univ Manitoba, Winnipeg, MB, Canada
[3] Huawei Technol Canada, Markham, ON, Canada
[4] Concordia Univ, Montreal, PQ, Canada
基金
中国国家自然科学基金; 加拿大自然科学与工程研究理事会;
关键词
D O I
10.1109/WACV56688.2023.00620
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We consider a new problem of few-shot learning of compact models. Meta-learning is a popular approach for fewshot learning. Previous work in meta-learning typically assumes that the model architecture during meta-training is the same as the model architecture used for final deployment. In this paper, we challenge this basic assumption. For final deployment, we often need the model to be small. But small models usually do not have enough capacity to effectively adapt to new tasks. In the mean time, we often have access to the large dataset and extensive computing power during meta-training since meta-training is typically performed on a server. In this paper, we propose task-specific meta distillation that simultaneously learns two models in meta-learning: a large teacher model and a small student model. These two models are jointly learned during meta-training. Given a new task during meta-testing, the teacher model is first adapted to this task, then the adapted teacher model is used to guide the adaptation of the student model. The adapted student model is used for final deployment. We demonstrate the effectiveness of our approach in few-shot image classification using model-agnostic metal-earning (MAML). Our proposed method outperforms other alternatives on several benchmark datasets.
引用
收藏
页码:6254 / 6263
页数:10
相关论文
共 50 条
  • [41] Meta-Transfer Learning for Few-Shot Learning
    Sun, Qianru
    Liu, Yaoyao
    Chua, Tat-Seng
    Schiele, Bernt
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 403 - 412
  • [42] Few-shot code translation via task-adapted prompt learning
    Li, Xuan
    Yuan, Shuai
    Gu, Xiaodong
    Chen, Yuting
    Shen, Beijun
    JOURNAL OF SYSTEMS AND SOFTWARE, 2024, 212
  • [43] Knowledge Distillation Meets Few-Shot Learning: An Approach for Few-Shot Intent Classification Within and Across Domains
    Sauer, Anna
    Asaadi, Shima
    Kuech, Fabian
    PROCEEDINGS OF THE 4TH WORKSHOP ON NLP FOR CONVERSATIONAL AI, 2022, : 108 - 119
  • [44] Few-Shot Face Stylization via GAN Prior Distillation
    Zhao, Ruoyu
    Zhu, Mingrui
    Wang, Nannan
    Gao, Xinbo
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2025, 36 (03) : 4492 - 4503
  • [45] Few-Shot Face Stylization via GAN Prior Distillation
    Zhao, Ruoyu
    Zhu, Mingrui
    Wang, Nannan
    Gao, Xinbo
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, : 1 - 12
  • [46] True Few-Shot Learning with Language Models
    Perez, Ethan
    Kiela, Douwe
    Cho, Kyunghyun
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [47] Task-Equivariant Graph Few-shot Learning
    Kim, Sungwon
    Lee, Junseok
    Lee, Namkyeong
    Kim, Wonjoong
    Choi, Seungyoon
    Park, Chanyoung
    PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023, 2023, : 1120 - 1131
  • [48] DETA: Denoised Task Adaptation for Few-Shot Learning
    Zhang, Ji
    Gao, Lianli
    Luo, Xu
    Shen, Hengtao
    Song, Jingkuan
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 11507 - 11517
  • [49] Task Encoding With Distribution Calibration for Few-Shot Learning
    Zhang, Jing
    Zhang, Xinzhou
    Wang, Zhe
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (09) : 6240 - 6252
  • [50] Few-shot learning with adaptively initialized task optimizer: a practical meta-learning approach
    Ye, Han-Jia
    Sheng, Xiang-Rong
    Zhan, De-Chuan
    MACHINE LEARNING, 2020, 109 (03) : 643 - 664