Multi-task Self-Supervised Visual Learning

被引:354
|
作者
Doersch, Carl [1 ]
Zisserman, Andrew [1 ,2 ]
机构
[1] DeepMind, London, England
[2] Univ Oxford, Dept Engn Sci, VGG, Oxford, England
关键词
D O I
10.1109/ICCV.2017.226
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We investigate methods for combining multiple self-supervised tasks-i.e., supervised tasks where data can be collected without manual labeling-in order to train a single visual representation. First, we provide an apples-to-apples comparison of four different self-supervised tasks using the very deep ResNet-101 architecture. We then combine tasks to jointly train a network. We also explore lasso regularization to encourage the network to factorize the information in its representation, and methods for "harmonizing" network inputs in order to learn a more unified representation. We evaluate all methods on ImageNet classification, PASCAL VOC detection, and NYU depth prediction. Our results show that deeper networks work better, and that combining tasks-even via a naive multihead architecture-always improves performance. Our best joint network nearly matches the PASCAL performance of a model pre-trained on ImageNet classification, and matches the ImageNet network on NYU depth prediction.
引用
收藏
页码:2070 / 2079
页数:10
相关论文
共 50 条
  • [1] MULTI-TASK SELF-SUPERVISED VISUAL REPRESENTATION LEARNING FOR MONOCULAR ROAD SEGMENTATION
    Cho, Jaehoon
    Kim, Youngjung
    Jung, Hyungjoo
    Oh, Changjae
    Youn, Jaesung
    Sohn, Kwanghoon
    [J]. 2018 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2018,
  • [2] Multi-task Semantic Matching with Self-supervised Learning
    Chen, Yuan
    Qiu, Xinying
    [J]. Beijing Daxue Xuebao (Ziran Kexue Ban)/Acta Scientiarum Naturalium Universitatis Pekinensis, 2022, 58 (01): : 83 - 90
  • [3] Multi-task Self-Supervised Adaptation for Reinforcement Learning
    Wu, Keyu
    Chen, Zhenghua
    Wu, Min
    Xiang, Shili
    Jin, Ruibing
    Zhang, Le
    Li, Xiaoli
    [J]. 2022 IEEE 17TH CONFERENCE ON INDUSTRIAL ELECTRONICS AND APPLICATIONS (ICIEA), 2022, : 15 - 20
  • [4] Multi-Task Self-Supervised Learning for Disfluency Detection
    Wang, Shaolei
    Che, Wanxiang
    Liu, Qi
    Qin, Pengda
    Liu, Ting
    Wang, William Yang
    [J]. THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 9193 - 9200
  • [5] MULTI-TASK SELF-SUPERVISED LEARNING FOR ROBUST SPEECH RECOGNITION
    Ravanelli, Mirco
    Zhong, Jianyuan
    Pascual, Santiago
    Swietojanski, Pawel
    Monteiro, Joao
    Trmal, Jan
    Bengio, Yoshua
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 6989 - 6993
  • [6] A MULTI-TASK SELF-SUPERVISED LEARNING FRAMEWORK FOR SCOPY IMAGES
    Li, Yuexiang
    Chen, Jiawei
    Zheng, Yefeng
    [J]. 2020 IEEE 17TH INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING (ISBI 2020), 2020, : 2005 - 2009
  • [7] Multi-Task Self-Supervised Learning for Script Event Prediction
    Zhou, Bo
    Chen, Yubo
    Liu, Kang
    Zhao, Jun
    Xu, Jiexin
    Jiang, Xiaojian
    Li, Jinlong
    [J]. PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 3662 - 3666
  • [8] Self-supervised multi-task learning for medical image analysis
    Yu, Huihui
    Dai, Qun
    [J]. PATTERN RECOGNITION, 2024, 150
  • [9] Self-supervised Multi-task Representation Learning for Sequential Medical Images
    Dong, Nanqing
    Kampffmeyer, Michael
    Voiculescu, Irina
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2021: RESEARCH TRACK, PT III, 2021, 12977 : 779 - 794
  • [10] Learning Representations for Bipartite Graphs Using Multi-task Self-supervised Learning
    Sethi, Akshay
    Gupta, Sonia
    Malhotra, Aakarsh
    Asthana, Siddhartha
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, ECML PKDD 2023, PT III, 2023, 14171 : 19 - 35