Embedding Global Contrastive and Local Location in Self-Supervised Learning

被引:22
|
作者
Zhao, Wenyi [1 ]
Li, Chongyi [2 ]
Zhang, Weidong [3 ]
Yang, Lu [1 ]
Zhuang, Peixian [4 ]
Li, Lingqiao [5 ]
Fan, Kefeng [6 ]
Yang, Huihua [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Sch Artificial Intelligence, Beijing 100876, Peoples R China
[2] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore 639798, Singapore
[3] Henan Inst Sci & Technol, Sch Informat Engn, Xinxiang 453600, Peoples R China
[4] Univ Sci & Technol Beijing, Sch Automat & Elect Engn, Key Lab Knowledge Automat Ind Proc, Minist Educ, Beijing 100083, Peoples R China
[5] Guilin Univ Elect Technol, Sch Comp Sci & Informat Secur, Guilin 541004, Peoples R China
[6] Chinese Elect Standardizat Inst, Beijing 101102, Peoples R China
基金
中国国家自然科学基金;
关键词
Task analysis; Optimization; Feature extraction; Semantics; Training; Ensemble learning; Data models; Self-supervised representation learning; contrastive learning; location-based sampling; ensemble learning; MOMENTUM CONTRAST;
D O I
10.1109/TCSVT.2022.3221611
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Self-supervised representation learning (SSL) typically suffers from inadequate data utilization and feature-specificity due to the suboptimal sampling strategy and the monotonous optimization method. Existing contrastive-based methods alleviate these issues through exceedingly long training time and large batch size, resulting in non-negligible computational consumption and memory usage. In this paper, we present an efficient self-supervised framework, called GLNet. The key insights of this work are the novel sampling and ensemble learning strategies embedded in the self-supervised framework. We first propose a location-based sampling strategy to integrate the complementary advantages of semantic and spatial characteristics. Whereafter, a Siamese network with momentum update is introduced to generate representative vectors, which are used to optimize the feature extractor. Finally, we particularly embed global contrastive and local location tasks in the framework, which aims to leverage the complementarity between the high-level semantic features and low-level texture features. Such complementarity is significant for mitigating the feature-specificity and improving the generalizability, thus effectively improving the performance of downstream tasks. Extensive experiments on representative benchmark datasets demonstrate that GLNet performs favorably against the state-of-the-art SSL methods. Specifically, GLNet improves MoCo-v3 by 2.4% accuracy on ImageNet dataset, while improves 2% accuracy and consumes only 75% training time on the ImageNet-100 dataset. In addition, GLNet is appealing in its compatibility with popular SSL frameworks. Code is available at GLNet.
引用
收藏
页码:2275 / 2289
页数:15
相关论文
共 50 条
  • [41] Interactive Contrastive Learning for Self-Supervised Entity Alignment
    Zeng, Kaisheng
    Dong, Zhenhao
    Hou, Lei
    Cao, Yixin
    Hu, Minghao
    Yu, Jifan
    Lv, Xin
    Cao, Lei
    Wang, Xin
    Liu, Haozhuang
    Huang, Yi
    Feng, Junlan
    Wan, Jing
    Li, Juanzi
    Feng, Ling
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 2465 - 2475
  • [42] Grouped Contrastive Learning of Self-Supervised Sentence Representation
    Wang, Qian
    Zhang, Weiqi
    Lei, Tianyi
    Peng, Dezhong
    APPLIED SCIENCES-BASEL, 2023, 13 (17):
  • [43] Contrastive self-supervised learning for neurodegenerative disorder classification
    Gryshchuk, Vadym
    Singh, Devesh
    Teipel, Stefan
    Dyrba, Martin
    ADNI Study Grp
    AIBL Study Grp
    FTLDNI Study Grp
    FRONTIERS IN NEUROINFORMATICS, 2025, 19
  • [44] Contrastive Self-Supervised Learning for Optical Music Recognition
    Penarrubia, Carlos
    Valero-Mas, Jose J.
    Calvo-Zaragoza, Jorge
    DOCUMENT ANALYSIS SYSTEMS, DAS 2024, 2024, 14994 : 312 - 326
  • [45] Contrastive UCB: Provably Efficient Contrastive Self-Supervised Learning in Online Reinforcement Learning
    Qiu, Shuang
    Wang, Lingxiao
    Bai, Chenjia
    Yang, Zhuoran
    Wang, Zhaoran
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [46] GLOCAL: A self-supervised learning framework for global and local motion estimation
    Zheng, Yihao
    Luo, Kunming
    Liu, Shuaicheng
    Li, Zun
    Xiang, Ye
    Wu, Lifang
    Zeng, Bing
    Chen, Chang Wen
    PATTERN RECOGNITION LETTERS, 2024, 178 : 91 - 97
  • [47] Bayesian Self-Supervised Learning Using Local and Global Graph Information
    Polyzos, Konstantinos D.
    Sadeghi, Alireza
    Giannakis, Georgios B.
    2023 IEEE 9TH INTERNATIONAL WORKSHOP ON COMPUTATIONAL ADVANCES IN MULTI-SENSOR ADAPTIVE PROCESSING, CAMSAP, 2023, : 256 - 260
  • [48] Similarity contrastive estimation for image and video soft contrastive self-supervised learning
    Denize, Julien
    Rabarisoa, Jaonary
    Orcesi, Astrid
    Herault, Romain
    MACHINE VISION AND APPLICATIONS, 2023, 34 (06)
  • [49] Investigating Contrastive Pair Learning's Frontiers in Supervised, Semisupervised, and Self-Supervised Learning
    Sabiri, Bihi
    Khtira, Amal
    EL Asri, Bouchra
    Rhanoui, Maryem
    JOURNAL OF IMAGING, 2024, 10 (08)
  • [50] Similarity contrastive estimation for image and video soft contrastive self-supervised learning
    Julien Denize
    Jaonary Rabarisoa
    Astrid Orcesi
    Romain Hérault
    Machine Vision and Applications, 2023, 34