Positive-Unlabeled Learning for Knowledge Distillation

被引:0
|
作者
Ning Jiang
Jialiang Tang
Wenxin Yu
机构
[1] Southwest University of Science and Technology,School of Computer Science and Technology
来源
Neural Processing Letters | 2023年 / 55卷
关键词
Convolutional neural networks; Model compression; Knowledge distillation; Positive-unlabeled learning; Attention mechanism; Soft-target;
D O I
暂无
中图分类号
学科分类号
摘要
Convolutional neural networks (CNNs) have greatly promoted the development of artificial intelligence. In general, CNNs with high performance are over-parameterized, requiring massive calculations to process and predict the data. It leads CNNs unable to apply to exiting resource-limited intelligence devices. In this paper, we propose an efficient model compression framework based on knowledge distillation to train a compact student network by a large teacher network. Our key point is to introduce a positive-unlabeled (PU) classifier to promote the compressed student network to learn the features of the prominent teacher network as much as possible. During the training, the PU classifier is to discriminate the features of the teacher network as high-quality and discriminate the features of the student network as low-quality. Simultaneously, the student network learns knowledge from the teacher network through the soft-targets and attention features. Extensive experimental evaluations on four benchmark image classification datasets show that our method outperforms the prior works with a large margin at the same parameters and calculations cost. When selecting the VGGNet19 as the teacher network to train on the CIFAR dataset, the student network VGGNet13 achieves 94.47% and 75.73% accuracy on the CIFAR-10 and CIFAR-100 datasets, which improved 1.02% and 2.44%, respectively.
引用
收藏
页码:2613 / 2631
页数:18
相关论文
共 50 条
  • [1] Positive-Unlabeled Learning for Knowledge Distillation
    Jiang, Ning
    Tang, Jialiang
    Yu, Wenxin
    NEURAL PROCESSING LETTERS, 2023, 55 (03) : 2613 - 2631
  • [2] Knowledge Distillation Based on Positive-Unlabeled Classification and Attention Mechanism
    Tang, Jialiang
    Liu, Mingjin
    Jiang, Ning
    Yu, Wenxin
    Yang, Changzheng
    Zhou, Jinjia
    2021 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2021,
  • [3] Density Estimators for Positive-Unlabeled Learning
    Basile, Teresa M. A.
    Di Mauro, Nicola
    Esposito, Floriana
    Ferilli, Stefano
    Vergari, Antonio
    NEW FRONTIERS IN MINING COMPLEX PATTERNS, NFMCP 2017, 2018, 10785 : 49 - 64
  • [4] Generative Adversarial Positive-Unlabeled Learning
    Hou, Ming
    Chaib-draa, Brahim
    Li, Chao
    Zhao, Qibin
    PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2018, : 2255 - 2261
  • [5] Positive-Unlabeled Learning in Streaming Networks
    Chang, Shiyu
    Zhang, Yang
    Tang, Jiliang
    Yin, Dawei
    Chang, Yi
    Hasegawa-Johnson, Mark A.
    Huang, Thomas S.
    KDD'16: PROCEEDINGS OF THE 22ND ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2016, : 755 - 764
  • [6] A boosting framework for positive-unlabeled learning
    Zhao, Yawen
    Zhang, Mingzhe
    Zhang, Chenhao
    Chen, Weitong
    Ye, Nan
    Xu, Miao
    STATISTICS AND COMPUTING, 2025, 35 (01)
  • [7] Positive-Unlabeled Learning With Label Distribution Alignment
    Jiang, Yangbangyan
    Xu, Qianqian
    Zhao, Yunrui
    Yang, Zhiyong
    Wen, Peisong
    Cao, Xiaochun
    Huang, Qingming
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (12) : 15345 - 15363
  • [8] Positive-Unlabeled Learning for Network Link Prediction
    Gan, Shengfeng
    Alshahrani, Mohammed
    Liu, Shichao
    MATHEMATICS, 2022, 10 (18)
  • [9] Positive-Unlabeled Learning from Imbalanced Data
    Su, Guangxin
    Chen, Weitong
    Xu, Miao
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 2995 - 3001
  • [10] Ensembles of density estimators for positive-unlabeled learning
    T. M. A. Basile
    N. Di Mauro
    F. Esposito
    S. Ferilli
    A. Vergari
    Journal of Intelligent Information Systems, 2019, 53 : 199 - 217