Unsupervised Pre-Training of Image Features on Non-Curated Data

被引:143
|
作者
Caron, Mathilde [1 ,2 ]
Bojanowski, Piotr [1 ]
Mairal, Julien [2 ]
Joulin, Armand [1 ]
机构
[1] Facebook AI Res, Menlo Pk, CA 94025 USA
[2] Univ Grenoble Alpes, CNRS, INRIA, Grenoble INP,LJK, F-38000 Grenoble, France
基金
欧洲研究理事会;
关键词
D O I
10.1109/ICCV.2019.00305
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pre-training general-purpose visual features with convolutional neural networks without relying on annotations is a challenging and important task. Most recent efforts in unsupervised feature learning have focused on either small or highly curated datasets like ImageNet, whereas using non-curated raw datasets was found to decrease the feature quality when evaluated on a transfer task. Our goal is to bridge the performance gap between unsupervised methods trained on curated data, which are costly to obtain, and massive raw datasets that are easily available. To that effect, we propose a new unsupervised approach which leverages self-supervision and clustering to capture complementary statistics from large-scale data. We validate our approach on 96 million images from YFCC100M [42], achieving state-of-the-art results among unsupervised methods on standard benchmarks, which confirms the potential of unsupervised learning when only non-curated raw data are available. We also show that pre-training a supervised VGG-16 with our method achieves 74.9% top-1 classification accuracy on the validation set of ImageNet, which is an improvement of +0.8% over the same network trained from scratch. Our code is available at https://github.com/facebookresearch/DeeperCluster.
引用
收藏
页码:2959 / 2968
页数:10
相关论文
共 50 条
  • [41] Unsupervised Pre-Training for 3D Leaf Instance Segmentation
    Roggiolani, Gianmarco
    Magistri, Federico
    Guadagnino, Tiziano
    Behley, Jens
    Stachniss, Cyrill
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (11) : 7448 - 7455
  • [42] Image-Text Pre-Training for Logo Recognition
    Hubenthal, Mark
    Kumar, Suren
    [J]. 2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, : 1145 - 1154
  • [43] Explicit Cross-lingual Pre-training for Unsupervised Machine Translation
    Ren, Shuo
    Wu, Yu
    Liu, Shujie
    Zhou, Ming
    Ma, Shuai
    [J]. 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 770 - 779
  • [44] Image Difference Captioning with Pre-training and Contrastive Learning
    Yao, Linli
    Wang, Weiying
    Jin, Qin
    [J]. THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 3108 - 3116
  • [45] Data Determines Distributional Robustness in Contrastive Language-Image Pre-training (CLIP)
    Fang, Alex
    Ilharco, Gabriel
    Wortsman, Mitchell
    Wan, Yuhao
    Shankar, Vaishaal
    Dave, Achal
    Schmidt, Ludwig
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [46] Data Labeling Using Unsupervised Cascaded Pre-training with Fused Multi-port Data for Optical Failure Management
    Yang, Weijie
    Zhang, Chunyu
    Wang, Danshi
    Zhu, Hong
    Xu, Xinxing
    Shi, Degang
    Zhang, Min
    [J]. 2024 OPTICAL FIBER COMMUNICATIONS CONFERENCE AND EXHIBITION, OFC, 2024,
  • [47] Deeply Unsupervised Patch Re-Identification for Pre-Training Object Detectors
    Ding, Jian
    Xie, Enze
    Xu, Hang
    Jiang, Chenhan
    Li, Zhenguo
    Luo, Ping
    Xia, Gui-Song
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (03) : 1348 - 1361
  • [48] Unsupervised bin-wise pre-training: A fusion of information theory and hypergraph
    Glory, Anila H.
    Vigneswaran, C.
    Sriram, Shankar V. S.
    [J]. KNOWLEDGE-BASED SYSTEMS, 2020, 195
  • [49] Application Specificity of Data for Pre-Training in Computer Vision
    Peters, Gabriel G.
    Couwenhoven, Scott D.
    Walvoord, Derek J.
    Salvaggio, Carl
    [J]. DISRUPTIVE TECHNOLOGIES IN INFORMATION SCIENCES VIII, 2024, 13058
  • [50] Lottery Hypothesis based Unsupervised Pre-training for Model Compression in Federated Learning
    Itahara, Sohei
    Nishio, Takayuki
    Morikura, Masahiro
    Yamamoto, Koji
    [J]. 2020 IEEE 92ND VEHICULAR TECHNOLOGY CONFERENCE (VTC2020-FALL), 2020,