Knowledge Distillation Meets Open-Set Semi-supervised Learning

被引:0
|
作者
Yang, Jing [1 ]
Zhu, Xiatian [2 ,3 ]
Bulat, Adrian [2 ]
Martinez, Brais [2 ]
Tzimiropoulos, Georgios [2 ,4 ]
机构
[1] Univ Nottingham, Nottingham, England
[2] Samsung AI Ctr, Cambridge, England
[3] Univ Surrey, Guildford, England
[4] Queen Mary Univ London, London, England
关键词
Knowledge distillation; Structured representational knowledge; Open-set semi-supervised learning; Out-of-distribution;
D O I
10.1007/s11263-024-02192-7
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Existing knowledge distillation methods mostly focus on distillation of teacher's prediction and intermediate activation. However, the structured representation, which arguably is one of the most critical ingredients of deep models, is largely overlooked. In this work, we propose a novel semantic representational distillation (SRD) method dedicated for distilling representational knowledge semantically from a pretrained teacher to a target student. The key idea is that we leverage the teacher's classifier as a semantic critic for evaluating the representations of both teacher and student and distilling the semantic knowledge with high-order structured information over all feature dimensions. This is accomplished by introducing a notion of cross-network logit computed through passing student's representation into teacher's classifier. Further, considering the set of seen classes as a basis for the semantic space in a combinatorial perspective, we scale SRD to unseen classes for enabling effective exploitation of largely available, arbitrary unlabeled training data. At the problem level, this establishes an interesting connection between knowledge distillation with open-set semi-supervised learning (SSL). Extensive experiments show that our SRD outperforms significantly previous state-of-the-art knowledge distillation methods on both coarse object classification and fine face recognition tasks, as well as less studied yet practically crucial binary network distillation. Under more realistic open-set SSL settings we introduce, we reveal that knowledge distillation is generally more effective than existing out-of-distribution sample detection, and our proposed SRD is superior over both previous distillation and SSL competitors. The source code is available at https://github.com/jingyang2017/SRD_ossl.
引用
收藏
页码:315 / 334
页数:20
相关论文
共 50 条
  • [41] Self Pseudo Entropy Knowledge Distillation for Semi-supervised Semantic Segmentation
    Lu X.
    Jiao L.
    Li L.
    Liu F.
    Liu X.
    Yang S.
    IEEE Trans Circuits Syst Video Technol, 8 (7359-7372): : 1 - 1
  • [42] Expert-Guided Knowledge Distillation for Semi-Supervised Vessel Segmentation
    Shen, Ning
    Xu, Tingfa
    Huang, Shiqi
    Mu, Feng
    Li, Jianan
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2023, 27 (11) : 5542 - 5553
  • [43] Semi-supervised Learning Meets Factorization: Learning to Recommend with Chain Graph Model
    Chen, Chaochao
    Chang, Kevin Chen-Chuan
    Li, Qibing
    Zheng, Xiaolin
    ACM TRANSACTIONS ON KNOWLEDGE DISCOVERY FROM DATA, 2018, 12 (06)
  • [44] Open problems in efficient semi-supervised PAC learning
    Blum, Avrim
    Balcan, Maria-Florina
    LEARNING THEORY, PROCEEDINGS, 2007, 4539 : 622 - +
  • [45] Collaborative Graph Convolutional Networks: Unsupervised Learning Meets Semi-Supervised Learning
    Hui, Binyuan
    Zhu, Pengfei
    Hu, Qinghua
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 4215 - 4222
  • [46] On semi-supervised learning
    A. Cholaquidis
    R. Fraiman
    M. Sued
    TEST, 2020, 29 : 914 - 937
  • [47] On semi-supervised learning
    Cholaquidis, A.
    Fraiman, R.
    Sued, M.
    TEST, 2020, 29 (04) : 914 - 937
  • [48] Semi-supervised Learning
    Adams, Niall
    JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES A-STATISTICS IN SOCIETY, 2009, 172 : 530 - 530
  • [49] Fuzzy-Rough Set based Semi-Supervised Learning
    Mac Parthalain, Neil
    Jensen, Richard
    IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS (FUZZ 2011), 2011, : 2465 - 2472
  • [50] Learning Bounds for Open-Set Learning
    Fang, Zhen
    Lu, Jie
    Liu, Anjin
    Liu, Feng
    Zhang, Guangquan
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139