Repdistiller: Knowledge Distillation Scaled by Re-parameterization for Crowd Counting

被引:0
|
作者
Ni, Tian [1 ]
Cao, Yuchen [1 ]
Liang, Xiaoyu [1 ]
Hu, Haoji [1 ]
机构
[1] Zhejiang Univ, Coll Informat Sci & Elect Engn, Hangzhou, Peoples R China
关键词
Crowd counting; Knowledge distillation; Structural re-parameterization;
D O I
10.1007/978-981-99-8549-4_32
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Knowledge distillation (KD) is an important method to compress a large teacher model into a much smaller student model. However, the large capacity gap between the teacher and student models hinders the performance of KD in various tasks. In this paper, we propose Repdistiller, a knowledge distillation framework combined with structural re-parameterization to alleviate the capacity gap problem. Repdistiller makes the student model search for parallel branches during training, thus the capacity gap between the teacher and student models is decreased. After knowledge distillation, the searched branches are merged into the student network without causing any computation overhead for inference. Taking the crowd counting task as an example, Repdistiller achieves state-of-the-art performance on the ShanghaiTech and UCF-QNRF datasets, outperforming many well-established knowledge distillation methods.
引用
收藏
页码:383 / 394
页数:12
相关论文
共 50 条
  • [1] Online Convolutional Re-parameterization
    Hu, Mu
    Feng, Junyi
    Hua, Jiashen
    Lai, Baisheng
    Huang, Jianqiang
    Gong, Xiaojin
    Hua, Xiansheng
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 558 - 567
  • [2] Efficient Crowd Counting via Dual Knowledge Distillation
    Wang, Rui
    Hao, Yixue
    Hu, Long
    Li, Xianzhi
    Chen, Min
    Miao, Yiming
    Humar, Iztok
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 569 - 583
  • [3] Improved Knowledge Distillation for Crowd Counting on IoT Devices
    Huang, Zuo
    Sinnott, Richard O.
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON EDGE COMPUTING AND COMMUNICATIONS, EDGE, 2023, : 207 - 214
  • [4] A Re-parameterization Transformation of Bezier Curve
    Guo, Fenghua
    [J]. ADVANCES IN CIVIL AND INDUSTRIAL ENGINEERING, PTS 1-4, 2013, 353-356 : 3645 - 3648
  • [5] SHUFFLECOUNT: TASK-SPECIFIC KNOWLEDGE DISTILLATION FOR CROWD COUNTING
    Jiang, Minyang
    Lin, Jianzhe
    Wang, Z. Jane
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2021, : 999 - 1003
  • [6] Re-parameterization of multinomial distributions and diversity indices
    Zhang, Zhiyi
    Zhou, Jun
    [J]. JOURNAL OF STATISTICAL PLANNING AND INFERENCE, 2010, 140 (07) : 1731 - 1738
  • [7] Re-parameterization invariance in fractional flux periodicity
    Murakami, S
    Sasaki, K
    Saito, R
    [J]. JOURNAL OF THE PHYSICAL SOCIETY OF JAPAN, 2004, 73 (12) : 3231 - 3234
  • [8] DyRep: Bootstrapping Training with Dynamic Re-parameterization
    Huang, Tao
    You, Shan
    Zhang, Bohan
    Du, Yuxuan
    Wang, Fei
    Qian, Chen
    Xu, Chang
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 578 - 587
  • [9] Bayesian based Re-parameterization for DNN Model Pruning
    Lu, Xiaotong
    Xi, Teng
    Li, Baopu
    Zhang, Gang
    Dong, Weisheng
    Shi, Guangming
    [J]. PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 1367 - 1375
  • [10] Model re-parameterization and output prediction for a bioreactor system
    Surisetty, Kartik
    Siegler, Hector De la Hoz
    McCaffrey, WilliamC.
    Ben-Zvi, Amos
    [J]. CHEMICAL ENGINEERING SCIENCE, 2010, 65 (16) : 4535 - 4547