Cross-Architecture Distillation for Face Recognition

被引:0
|
作者
Zhao, Weisong [1 ]
Zhu, Xiangyu [2 ,3 ]
He, Zhixiang [4 ]
Zhang, Xiao-Yu [1 ,5 ]
Lei, Zhen [2 ,3 ,6 ]
机构
[1] Chinese Acad Sci, Inst Informat Engn, Beijing, Peoples R China
[2] UCAS, Sch Artificial Intelligence, Beijing, Peoples R China
[3] CASIA, MAIS, Beijing, Peoples R China
[4] Data&AI Technol Co, China Telecom Corp Ltd, Beijing, Peoples R China
[5] UCAS, Sch Cyber Secur, Beijing, Peoples R China
[6] Chinese Acad Sci, HKISI, CAIR, Beijing, Peoples R China
基金
北京市自然科学基金;
关键词
Face Recognition; Knowledge Distillation; Transformer;
D O I
10.1145/3581783.3611711
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Transformers have emerged as the superior choice for face recognition tasks, but their insufficient platform acceleration hinders their application on mobile devices. In contrast, Convolutional Neural Networks (CNNs) capitalize on hardware-compatible acceleration libraries. Consequently, it has become indispensable to preserve the distillation efficacy when transferring knowledge from a Transformer-based teacher model to a CNN-based student model, known as Cross-Architecture Knowledge Distillation (CAKD). Despite its potential, the deployment of CAKD in face recognition encounters two challenges: 1) the teacher and student share disparate spatial information for each pixel, obstructing the alignment of feature space, and 2) the teacher network is not trained in the role of a teacher, lacking proficiency in handling distillation-specific knowledge. To surmount these two constraints, 1) we first introduce a Unified Receptive Fields Mapping module (URFM) that maps pixel features of the teacher and student into local features with unified receptive fields, thereby synchronizing the pixel-wise spatial information of teacher and student. Subsequently, 2) we develop an Adaptable Prompting Teacher network (APT) that integrates prompts into the teacher, enabling it to manage distillation-specific knowledge while preserving the model's discriminative capacity. Extensive experiments on popular face benchmarks and two large-scale verification sets demonstrate the superiority of our method.
引用
收藏
页码:8076 / 8085
页数:10
相关论文
共 50 条
  • [1] Cross-Architecture Knowledge Distillation
    Liu, Yufan
    Cao, Jiajiong
    Li, Bing
    Hu, Weiming
    Ding, Jingting
    Li, Liang
    Maybank, Stephen
    [J]. INTERNATIONAL JOURNAL OF COMPUTER VISION, 2024, 132 (08) : 2798 - 2824
  • [2] Cross-Architecture Knowledge Distillation
    Liu, Yufan
    Cao, Jiajiong
    Li, Bing
    Hu, Weiming
    Ding, Jingting
    Li, Liang
    [J]. COMPUTER VISION - ACCV 2022, PT V, 2023, 13845 : 179 - 195
  • [3] Adaptive Cross-architecture Mutual Knowledge Distillation
    Ni, Jianyuan
    Tang, Hao
    Shang, Yuzhang
    Duan, Bin
    Yan, Yan
    [J]. 2024 IEEE 18TH INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION, FG 2024, 2024,
  • [4] General Cross-Architecture Distillation of Pretrained Language Models into Matrix Embeddings
    Galke, Lukas
    Cuber, Isabelle
    Meyer, Christoph
    Noelscher, Henrik Ferdinand
    Sonderecker, Angelina
    Scherp, Ansgar
    [J]. 2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [5] Cross-Architecture Lifter Synthesis
    van Tonder, Rijnard
    Le Goues, Claire
    [J]. SOFTWARE ENGINEERING AND FORMAL METHODS, SEFM 2018, 2018, 10886 : 155 - 170
  • [6] Cross-Architecture Bug Search in Binary Executables
    Pewny, Jannik
    Garmany, Behrad
    Gawlik, Robert
    Rossow, Christian
    Holz, Thorsten
    [J]. 2015 IEEE SYMPOSIUM ON SECURITY AND PRIVACY SP 2015, 2015, : 709 - 724
  • [7] Predicting Cross-Architecture Performance of Parallel Programs
    Nichols, Daniel
    Movsesyan, Alexander
    Yeom, Jae-Seung
    Sarkar, Abhik
    Milroy, Daniel
    Patki, Tapasya
    Bhatele, Abhinav
    [J]. PROCEEDINGS 2024 IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM, IPDPS 2024, 2024, : 570 - 581
  • [8] BinGo: Cross-Architecture Cross-OS Binary Search
    Chandramohan, Mahinthan
    Xue, Yinxing
    Xu, Zhengzi
    Liu, Yang
    Cho, Chia Yuan
    Kuan, Tan Hee Beng
    [J]. FSE'16: PROCEEDINGS OF THE 2016 24TH ACM SIGSOFT INTERNATIONAL SYMPOSIUM ON FOUNDATIONS OF SOFTWARE ENGINEERING, 2016, : 678 - 689
  • [9] DnD: A Cross-Architecture Deep Neural Network Decompiler
    Wu, Ruoyu
    Kim, Taegyu
    Tian, Dave
    Bianchi, Antonio
    Xu, Dongyan
    [J]. PROCEEDINGS OF THE 31ST USENIX SECURITY SYMPOSIUM, 2022, : 2135 - 2152
  • [10] Massive Similar Function Searching for Cross-Architecture Binaries
    Wang, Minhao
    Tang, Yong
    Lu, Zexin
    [J]. 2019 18TH INTERNATIONAL SYMPOSIUM ON DISTRIBUTED COMPUTING AND APPLICATIONS FOR BUSINESS ENGINEERING AND SCIENCE (DCABES 2019), 2019, : 195 - 198