Computation-Efficient Knowledge Distillation via Uncertainty-Aware Mixup

被引:17
|
作者
Xu, Guodong [1 ]
Liu, Ziwei [2 ]
Loy, Chen Change [2 ]
机构
[1] Chinese Univ Hong Kong, Hong Kong, Peoples R China
[2] Nanyang Technol Univ, Singapore, Singapore
关键词
Knowledge distillation; Training cost;
D O I
10.1016/j.patcog.2023.109338
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Knowledge distillation (KD) has emerged as an essential technique not only for model compression, but also other learning tasks such as continual learning. Given the richer application spectrum and potential online usage of KD, knowledge distillation efficiency becomes a pivotal component. In this work, we study this little-explored but important topic. Unlike previous works that focus solely on the accuracy of stu-dent network, we attempt to achieve a harder goal - to obtain a performance comparable to conventional KD with a lower computation cost during the transfer. To this end, we present UNcertainty-aware mIXup (UNIX), an effective approach that can reduce transfer cost by 20% to 30% and yet maintain comparable or achieve even better student performance than conventional KD. This is made possible via effective uncertainty sampling and a novel adaptive mixup approach that select informative samples dynamically over ample data and compact knowledge in these samples. We show that our approach inherently per-forms hard sample mining. We demonstrate the applicability of our approach to improve various existing KD approaches by reducing their queries to a teacher network. Extensive experiments are performed on CIFAR100 and ImageNet. Code and model are available at https://github.com/xuguodong03/UNIXKD .(c) 2023 Elsevier Ltd. All rights reserved.
引用
收藏
页数:9
相关论文
共 50 条
  • [1] UMIX: Improving ImportanceWeighting for Subpopulation Shift via Uncertainty-Aware Mixup
    Han, Zongbo
    Liang, Zhipeng
    Yang, Fan
    Liu, Liu
    Li, Lanqing
    Bian, Yatao
    Zhao, Peilin
    Wu, Bingzhe
    Zhang, Changqing
    Yao, Jianhua
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [2] Uncertainty-Aware Knowledge Distillation for Collision Identification of Collaborative Robots
    Kwon, Wookyong
    Jin, Yongsik
    Lee, Sang Jun
    SENSORS, 2021, 21 (19)
  • [3] Uncertainty-Aware Topological Persistence Guided Knowledge Distillation on Wearable Sensor Data
    Jeon, Eun Som
    Buman, Matthew P.
    Turaga, Pavan
    IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (18): : 30413 - 30429
  • [4] Real-Time Uncertainty Estimation in Computer Vision via Uncertainty-Aware Distribution Distillation
    Shen, Yichen
    Zhang, Zhilu
    Sabuncu, Mert R.
    Sun, Lin
    2021 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2021), 2021, : 707 - 716
  • [5] Uncertainty-Aware Contrastive Distillation for Incremental Semantic Segmentation
    Yang, Guanglei
    Fini, Enrico
    Xu, Dan
    Rota, Paolo
    Ding, Mingli
    Nabi, Moin
    Alameda-Pineda, Xavier
    Ricci, Elisa
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (02) : 2567 - 2581
  • [6] Robustness via Uncertainty-aware Cycle Consistency
    Upadhyay, Uddeshya
    Chen, Yanbei
    Akata, Zeynep
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [7] Robust Tracking via Uncertainty-Aware Semantic Consistency
    Ma, Jie
    Lan, Xiangyuan
    Zhong, Bineng
    Li, Guorong
    Tang, Zhenjun
    Li, Xianxian
    Ji, Rongrong
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (04) : 1740 - 1751
  • [8] Uncertainty-Aware Multi-Shot Knowledge Distillation for Image-Based Object Re-Identification
    Jin, Xin
    Lan, Cuiling
    Zeng, Wenjun
    Chen, Zhibo
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 11165 - 11172
  • [9] Uncertainty-Aware Prognosis via Deep Gaussian Process
    Biggio, Luca
    Wieland, Alexander
    Chao, Manuel Arias
    Kastanis, Iason
    Fink, Olga
    IEEE ACCESS, 2021, 9 : 123517 - 123527
  • [10] Predicting the Unpredictable: Uncertainty-Aware Reasoning over Temporal Knowledge Graphs via Diffusion Process
    Cai, Yuxiang
    Liu, Qiao
    Gan, Yanglei
    Li, Changlin
    Liu, Xueyi
    Lin, Run
    Lu, Da
    Yang, Jiaye
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 5766 - 5778