Positive-Unlabeled Learning for Knowledge Distillation

被引:0
|
作者
Ning Jiang
Jialiang Tang
Wenxin Yu
机构
[1] Southwest University of Science and Technology,School of Computer Science and Technology
来源
Neural Processing Letters | 2023年 / 55卷
关键词
Convolutional neural networks; Model compression; Knowledge distillation; Positive-unlabeled learning; Attention mechanism; Soft-target;
D O I
暂无
中图分类号
学科分类号
摘要
Convolutional neural networks (CNNs) have greatly promoted the development of artificial intelligence. In general, CNNs with high performance are over-parameterized, requiring massive calculations to process and predict the data. It leads CNNs unable to apply to exiting resource-limited intelligence devices. In this paper, we propose an efficient model compression framework based on knowledge distillation to train a compact student network by a large teacher network. Our key point is to introduce a positive-unlabeled (PU) classifier to promote the compressed student network to learn the features of the prominent teacher network as much as possible. During the training, the PU classifier is to discriminate the features of the teacher network as high-quality and discriminate the features of the student network as low-quality. Simultaneously, the student network learns knowledge from the teacher network through the soft-targets and attention features. Extensive experimental evaluations on four benchmark image classification datasets show that our method outperforms the prior works with a large margin at the same parameters and calculations cost. When selecting the VGGNet19 as the teacher network to train on the CIFAR dataset, the student network VGGNet13 achieves 94.47% and 75.73% accuracy on the CIFAR-10 and CIFAR-100 datasets, which improved 1.02% and 2.44%, respectively.
引用
收藏
页码:2613 / 2631
页数:18
相关论文
共 50 条
  • [21] Recovering True Classifier Performance in Positive-Unlabeled Learning
    Jain, Shantanu
    White, Martha
    Radivojac, Predrag
    THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 2066 - 2072
  • [22] Spotting Fake Reviews using Positive-Unlabeled Learning
    Li, Huayi
    Liu, Bing
    Mukherjee, Arjun
    Shao, Jidong
    COMPUTACION Y SISTEMAS, 2014, 18 (03): : 467 - 475
  • [23] Theoretical Comparisons of Positive-Unlabeled Learning against Positive-Negative Learning
    Niu, Gang
    du Plessis, Marthinus C.
    Sakai, Tomoya
    Ma, Yao
    Sugiyama, Masashi
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29
  • [24] Investigating Active Positive-Unlabeled Learning with Deep Networks
    Han, Kun
    Chen, Weitong
    Xu, Miao
    AI 2021: ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, 13151 : 607 - 618
  • [25] Bootstrap Latent Prototypes for graph positive-unlabeled learning
    Liang, Chunquan
    Tian, Yi
    Zhao, Dongmin
    Li, Mei
    Pan, Shirui
    Zhang, Hongming
    Wei, Jicheng
    INFORMATION FUSION, 2024, 112
  • [26] Positive-Unlabeled Compression on the Cloud
    Xu, Yixing
    Wang, Yunhe
    Chen, Hanting
    Han, Kai
    Xu, Chunjing
    Tao, Dacheng
    Xu, Chang
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [27] Positive-Unlabeled Domain Adaptation
    Sonntag, Jonas
    Behrens, Gunnar
    Schmidt-Thieme, Lars
    2022 IEEE 9TH INTERNATIONAL CONFERENCE ON DATA SCIENCE AND ADVANCED ANALYTICS (DSAA), 2022, : 66 - 75
  • [28] GradPU: Positive-Unlabeled Learning via Gradient Penalty and Positive Upweighting
    Dai, Songmin
    Li, Xiaoqiang
    Zhou, Yue
    Ye, Xichen
    Liu, Tong
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 6, 2023, : 7296 - +
  • [29] PUe: Biased Positive-Unlabeled Learning Enhancement by Causal Inference
    Wang, Xutao
    Chen, Hanting
    Guo, Tianyu
    Wang, Yunhe
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [30] PULNS: Positive-Unlabeled Learning with Effective Negative Sample Selector
    Luo, Chuan
    Zhao, Pu
    Chen, Chen
    Qiao, Bo
    Du, Chao
    Zhang, Hongyu
    Wu, Wei
    Cai, Shaowei
    He, Bing
    Rajmohan, Saravanakumar
    Lin, Qingwei
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 8784 - 8792