Towards a Smaller Student: Capacity Dynamic Distillation for Efficient Image Retrieval

被引:10
|
作者
Xie, Yi [1 ]
Zhang, Huaidong [1 ]
Xu, Xuemiao [1 ,4 ,5 ,6 ]
Zhu, Jianqing [2 ]
He, Shengfeng [3 ]
机构
[1] South China Univ Technol, Guangzhou, Guangdong, Peoples R China
[2] Huaqiao Univ, Quanzhou, Peoples R China
[3] Singapore Management Univ, Singapore, Singapore
[4] State Key Lab Subtrop Bldg Sci, Guangzhou, Guangdong, Peoples R China
[5] Minist Educ, Key Lab Big Data & Intelligent Robot, Guangzhou, Guangdong, Peoples R China
[6] Guangdong Prov Key Lab Computat Intelligence & Cy, Guangzhou, Guangdong, Peoples R China
基金
中国国家自然科学基金;
关键词
NETWORK;
D O I
10.1109/CVPR52729.2023.01536
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Previous Knowledge Distillation based efficient image retrieval methods employ a lightweight network as the student model for fast inference. However, the lightweight student model lacks adequate representation capacity for effective knowledge imitation during the most critical early training period, causing final performance degeneration. To tackle this issue, we propose a Capacity Dynamic Distillation framework, which constructs a student model with editable representation capacity. Specifically, the employed student model is initially a heavy model to fruitfully learn distilled knowledge in the early training epochs, and the student model is gradually compressed during the training. To dynamically adjust the model capacity, our dynamic framework inserts a learnable convolutional layer within each residual block in the student model as the channel importance indicator. The indicator is optimized simultaneously by the image retrieval loss and the compression loss, and a retrieval-guided gradient resetting mechanism is proposed to release the gradient conflict. Extensive experiments show that our method has superior inference speed and accuracy, e.g., on the VeRi-776 dataset, given the ResNet101 as a teacher, our method saves 67.13% model parameters and 65.67% FLOPs without sacrificing accuracy. Code is available at https://github.com/SCY-X/Capacity_Dynamic_Distillation.
引用
收藏
页码:16006 / 16015
页数:10
相关论文
共 50 条
  • [41] Unsupervised Deep Learning for Phase Retrieval via Teacher-Student Distillation
    Quan, Yuhui
    Chen, Zhile
    Pang, Tongyao
    Ji, Hui
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 2, 2023, : 2128 - 2136
  • [42] SmartAlbum -: Towards unification of approaches for image retrieval
    Tan, T
    Chen, J
    Mulhem, P
    16TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION, VOL III, PROCEEDINGS, 2002, : 983 - 986
  • [43] Towards Semantic Based Image Retrieval : A Review
    Wang, Hui Hui
    Mohamad, Dzulkifli
    Ismail, N. A.
    SECOND INTERNATIONAL CONFERENCE ON DIGITAL IMAGE PROCESSING, 2010, 7546
  • [44] Parameter-Efficient and Student-Friendly Knowledge Distillation
    Rao, Jun
    Meng, Xv
    Ding, Liang
    Qi, Shuhan
    Liu, Xuebo
    Zhang, Min
    Tao, Dacheng
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 4230 - 4241
  • [45] ItrievalKD: An Iterative Retrieval Framework Assisted with Knowledge Distillation for Noisy Text-to-Image Retrieval
    Liu, Zhen
    Zhu, Yongxin
    Gao, Zhujin
    Sheng, Xin
    Xu, Linli
    ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2023, PT III, 2023, 13937 : 257 - 268
  • [46] Towards Being Parameter-Efficient: A Stratified Sparsely Activated Transformer with Dynamic Capacity
    Xu, Haoran
    Elbayad, Maha
    Murray, Kenton
    Maillard, Jean
    Goswami, Vedanuj
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 12858 - 12870
  • [47] Towards Doubly Efficient Private Information Retrieval
    Canetti, Ran
    Holmgren, Justin
    Richelson, Silas
    THEORY OF CRYPTOGRAPHY, TCC 2017, PT II, 2017, 10678 : 694 - 726
  • [48] Self-supervised Image Hash Retrieval Based On Adversarial Distillation
    Feng, Ping
    Zhang, Hanyun
    2022 ASIA CONFERENCE ON ALGORITHMS, COMPUTING AND MACHINE LEARNING (CACML 2022), 2022, : 732 - 737
  • [49] A Novel Indexing and Image Annotation Structure for Efficient Image Retrieval
    B. Prasanthi
    Suresh Pabboju
    D. Vasumathi
    Arabian Journal for Science and Engineering, 2018, 43 : 4203 - 4213
  • [50] CAPACITY-ACHIEVING PRIVATE INFORMATION RETRIEVAL SCHEME WITH A SMALLER SUB-PACKETIZATION
    Zhang, Wenqin
    Zhou, Zhengchun
    Parampalli, Udaya
    Sidorenko, Vladimir
    ADVANCES IN MATHEMATICS OF COMMUNICATIONS, 2021, 15 (02) : 347 - 363