Momentum Contrastive Pruning

被引:1
|
作者
Pan, Siyuan [1 ]
Qin, Yiming [1 ]
Li, Tingyao [1 ]
Li, Xiaoshuang [1 ]
Hou, Liang [2 ]
机构
[1] Shanghai Jiao Tong Univ, Shanghai, Peoples R China
[2] Chinese Acad Sci, Inst Comp Technol, Beijing, Peoples R China
关键词
D O I
10.1109/CVPRW56347.2022.00298
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Momentum contrast [16] (MoCo) for unsupervised visual representation learning has a close performance to supervised learning, but it sometimes possesses excess parameters. Extracting a subnetwork from an over-parameterized unsupervised network without sacrificing performance is of particular interest to accelerate inference speed. Typical pruning methods are not applicable for MoCo, because in the fine-tune stage after pruning, the slow update of the momentum encoder will undermine the pretrained encoder. In this paper, we propose a Momentum Contrastive Pruning (MCP) method, which prunes the momentum encoder instead to obtain a momentum subnet. It maintains an unpruned momentum encoder as a smooth transition scheme to alleviate the representation gap between the encoder and momentum subnet. To fulfill the sparsity requirements of the encoder, alternating direction method of multipliers [40] (ADMM) is adopted. Experiments prove that our MCP method can obtain a momentum subnet that has almost equal performance as the over-parameterized MoCo when transferred to downstream tasks, meanwhile has much less parameters and float operations per second (FLOPs).
引用
收藏
页码:2646 / 2655
页数:10
相关论文
共 50 条
  • [21] MoCo4SRec: A momentum contrastive learning framework for sequential recommendation
    Wei, Zihan
    Wu, Ning
    Li, Fengxia
    Wang, Ke
    Zhang, Wei
    EXPERT SYSTEMS WITH APPLICATIONS, 2023, 223
  • [22] MCGCL:Adversarial attack on graph contrastive learning based on momentum gradient candidates
    Zhang, Qi
    Qin, Zhenkai
    Zhang, Yunjie
    PLOS ONE, 2024, 19 (06):
  • [23] From Dense to Sparse: Contrastive Pruning for Better Pre-trained Language Model Compression
    Xu, Runxin
    Luo, Fuli
    Wang, Chengyu
    Chang, Baobao
    Huang, Jun
    Huang, Songfang
    Huang, Fei
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 11547 - 11555
  • [24] ConVaT: A Variational Generative Transformer With Momentum Contrastive Learning for Hyperspectral Image Classification
    Liang, Miaomiao
    Liu, Zuo
    Dong, Jian
    Yu, Lingjuan
    Yu, Xiangchun
    Li, Jun
    Jiao, Licheng
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2024, 21 : 1 - 5
  • [25] mcBERT: Momentum Contrastive Learning with BERT for Zero-Shot Slot Filling
    Heo, Seong-Hwan
    Lee, WonKee
    Lee, Jong-Hyeok
    INTERSPEECH 2022, 2022, : 1243 - 1247
  • [26] Fast Filter Pruning via Coarse-to-Fine Neural Architecture Search and Contrastive Knowledge Transfer
    Lee, Seunghyun
    Song, Byung Cheol
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (07) : 9674 - 9685
  • [27] MOCOLNet: A Momentum Contrastive Learning Network for Multimodal Aspect-Level Sentiment Analysis
    Mu, Jie
    Nie, Feiping
    Wang, Wei
    Xu, Jian
    Zhang, Jing
    Liu, Han
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (12) : 8787 - 8800
  • [28] Augmented Skeleton Based Contrastive Action Learning with Momentum LSTM for Unsupervised Action Recognition
    Rao, Haocong
    Xu, Shihao
    Hu, Xiping
    Cheng, Jun
    Hu, Bin
    INFORMATION SCIENCES, 2021, 569 : 90 - 109
  • [29] Fast-MoCo: Boost Momentum-Based Contrastive Learning with Combinatorial Patches
    Ci, Yuanzheng
    Lin, Chen
    Bai, Lei
    Ouyang, Wanli
    COMPUTER VISION, ECCV 2022, PT XXVI, 2022, 13686 : 290 - 306
  • [30] Momentum memory contrastive learning for transfer-based few-shot classification
    Tian, Runliang
    Shi, Hongmei
    APPLIED INTELLIGENCE, 2023, 53 (01) : 864 - 878