Embedding Global Contrastive and Local Location in Self-Supervised Learning

被引:22
|
作者
Zhao, Wenyi [1 ]
Li, Chongyi [2 ]
Zhang, Weidong [3 ]
Yang, Lu [1 ]
Zhuang, Peixian [4 ]
Li, Lingqiao [5 ]
Fan, Kefeng [6 ]
Yang, Huihua [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Sch Artificial Intelligence, Beijing 100876, Peoples R China
[2] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore 639798, Singapore
[3] Henan Inst Sci & Technol, Sch Informat Engn, Xinxiang 453600, Peoples R China
[4] Univ Sci & Technol Beijing, Sch Automat & Elect Engn, Key Lab Knowledge Automat Ind Proc, Minist Educ, Beijing 100083, Peoples R China
[5] Guilin Univ Elect Technol, Sch Comp Sci & Informat Secur, Guilin 541004, Peoples R China
[6] Chinese Elect Standardizat Inst, Beijing 101102, Peoples R China
基金
中国国家自然科学基金;
关键词
Task analysis; Optimization; Feature extraction; Semantics; Training; Ensemble learning; Data models; Self-supervised representation learning; contrastive learning; location-based sampling; ensemble learning; MOMENTUM CONTRAST;
D O I
10.1109/TCSVT.2022.3221611
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Self-supervised representation learning (SSL) typically suffers from inadequate data utilization and feature-specificity due to the suboptimal sampling strategy and the monotonous optimization method. Existing contrastive-based methods alleviate these issues through exceedingly long training time and large batch size, resulting in non-negligible computational consumption and memory usage. In this paper, we present an efficient self-supervised framework, called GLNet. The key insights of this work are the novel sampling and ensemble learning strategies embedded in the self-supervised framework. We first propose a location-based sampling strategy to integrate the complementary advantages of semantic and spatial characteristics. Whereafter, a Siamese network with momentum update is introduced to generate representative vectors, which are used to optimize the feature extractor. Finally, we particularly embed global contrastive and local location tasks in the framework, which aims to leverage the complementarity between the high-level semantic features and low-level texture features. Such complementarity is significant for mitigating the feature-specificity and improving the generalizability, thus effectively improving the performance of downstream tasks. Extensive experiments on representative benchmark datasets demonstrate that GLNet performs favorably against the state-of-the-art SSL methods. Specifically, GLNet improves MoCo-v3 by 2.4% accuracy on ImageNet dataset, while improves 2% accuracy and consumes only 75% training time on the ImageNet-100 dataset. In addition, GLNet is appealing in its compatibility with popular SSL frameworks. Code is available at GLNet.
引用
收藏
页码:2275 / 2289
页数:15
相关论文
共 50 条
  • [1] Contrastive and Non-Contrastive Self-Supervised Learning Recover Global and Local Spectral Embedding Methods
    Balestriero, Randall
    LeCun, Yann
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [2] Global and Local Contrastive Learning for Self-Supervised Skeleton-Based Action Recognition
    Hu, Jinhua
    Hou, Yonghong
    Guo, Zihui
    Gao, Jiajun
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (11) : 10578 - 10589
  • [3] Adversarial Self-Supervised Contrastive Learning
    Kim, Minseon
    Tack, Jihoon
    Hwang, Sung Ju
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS (NEURIPS 2020), 2020, 33
  • [4] A Survey on Contrastive Self-Supervised Learning
    Jaiswal, Ashish
    Babu, Ashwin Ramesh
    Zadeh, Mohammad Zaki
    Banerjee, Debapriya
    Makedon, Fillia
    TECHNOLOGIES, 2021, 9 (01)
  • [5] Self-Supervised Learning: Generative or Contrastive
    Liu, Xiao
    Zhang, Fanjin
    Hou, Zhenyu
    Mian, Li
    Wang, Zhaoyu
    Zhang, Jing
    Tang, Jie
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (01) : 857 - 876
  • [6] Graph Anomaly Detection via Multiscale Contrastive Self-Supervised Learning From Local to Global
    Wang, Xiaofeng
    Lai, Shuaiming
    Zhu, Shuailei
    Chen, Yuntao
    Lv, Laishui
    Qi, Yuanyuan
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2024,
  • [7] Contrastive Self-Supervised Speaker Embedding With Sequential Disentanglement
    Tu, Youzhi
    Mak, Man-Wai
    Chien, Jen-Tzung
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 2704 - 2715
  • [8] Global and Local Contrastive Self-Supervised Learning for Semantic Segmentation of HR Remote Sensing Images
    Li, Haifeng
    Li, Yi
    Zhang, Guo
    Liu, Ruoyun
    Huang, Haozhe
    Zhu, Qing
    Tao, Chao
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
  • [9] Self-supervised Learning with Local Contrastive Loss for Detection and Semantic Segmentation
    Islam, Ashraful
    Lundell, Ben
    Sawhney, Harpreet
    Sinha, Sudipta N.
    Morales, Peter
    Radke, Richard J.
    2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, : 5613 - 5622
  • [10] A comprehensive perspective of contrastive self-supervised learning
    Songcan CHEN
    Chuanxing GENG
    Frontiers of Computer Science, 2021, (04) : 102 - 104