Deep Contrastive Learning: A Survey

被引:0
|
作者
Zhang C.-S. [1 ]
Chen J. [1 ]
Li Q.-L. [1 ]
Deng B.-Q. [1 ]
Wang J. [1 ]
Chen C.-G. [1 ]
机构
[1] Henan Key Lab of Big Data Analysis and Processing, Henan University, Kaifeng
来源
关键词
Contrastive learning; deep learning; feature extraction; metric learning; self-supervised learning;
D O I
10.16383/j.aas.c220421
中图分类号
学科分类号
摘要
In deep learning, it has been a crucial research concern on how to make use of the vast amount of unlabeled data to enhance the feature extraction capability of deep neural networks, for which contrastive learning is an effective approach. It has attracted significant research effort in the past few years, and a large number of contrastive learning methods have been proposed. In this paper, we survey recent advances and progress in contrastive learning in a comprehensive way. We first propose a new taxonomy for contrastive learning, in which we divide existing methods into 5 categories, including 1) sample pair construction methods, 2) image augmentation methods, 3) network architecture level methods, 4) loss function level methods, and 5) applications. Based on our proposed taxonomy, we systematically review the methods in each category, and analyze the characteristics and differences of representative methods. Moreover, we report and compare the performance of different contrastive learning methods on the benchmark datasets. We also retrospect the history of contrastive learning and discuss the differences and connections among contrastive learning, self-supervised learning, and metric learning. Finally, we discuss remaining issues and challenges in contrastive learning and outlook its future directions. © 2023 Science Press. All rights reserved.
引用
收藏
页码:15 / 39
页数:24
相关论文
共 109 条
  • [41] Kumar S, Haresh S, Ahmed A, Konin A, Zia M Z, Tran Q H., Unsupervised action segmentation by joint representation learning and online clustering, Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 20142-20153, (2022)
  • [42] Han T D, Xie W D, Zisserman A., Self-supervised co-training for video representation learning, Proceedings of the 34th International Conference on Neural Information Processing Systems, (2020)
  • [43] Wang H B, Xiao R, Li S, Et al., Contrastive Label Disambiguation for Partial Label Learning, Proceedings of the 10th International Conference on Learning Representations, (2022)
  • [44] Yang F, Wu K, Zhang S Y, Jiang G N, Liu Y, Zheng F, Et al., Class-aware contrastive semi-supervised learning, Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), (2022)
  • [45] Tian Y L, Krishnan D, Isola P., Contrastive multiview coding, Proceedings of the 16th European Conference on Computer Vision (ECCV), pp. 776-794, (2020)
  • [46] Rai N, Adeli E, Lee K H, Gaidon A, Niebles J C., CoCon: Cooperative-contrastive learning, Proceedings of the IEEE/ CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 3379-3388, (2021)
  • [47] Peng X Y, Wang K, Zhu Z, Wang M, You Y., Crafting better contrastive views for siamese representation learning, Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 16010-16019, (2022)
  • [48] Ding S R, Li M M, Yang T Y, Qian R, Xu H H, Chen Q Y, Et al., Motion-aware contrastive video representation learning via foreground-background merging, Proceedings of the IEEE/ CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 9706-9716, (2022)
  • [49] Li S, Gong K X, Liu C H, Wang Y L, Qiao F, Cheng X J., MetaSAug: Meta semantic augmentation for long-tailed visual recognition, Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5208-5217, (2021)
  • [50] Wang Y L, Pan X R, Song S J, Zhang H, Wu C, Huang G., Implicit semantic data augmentation for deep networks, Proceedings of the 33rd International Conference on Neural Information Processing Systems, (2019)