Kaizen: Practical self-supervised continual learning with continual fine-tuning

被引:0
|
作者
Tang, Chi Ian [1 ,2 ]
Qendrol, Lorena [1 ]
Spathis, Dimitris [1 ]
Kawsar, Fahim [1 ]
Mascolo, Cecilia [2 ]
Mathur, Akhil [1 ]
机构
[1] Nokia Bell Labs, Cambridge, England
[2] Univ Cambridge, Cambridge, England
关键词
D O I
10.1109/WACV57701.2024.00282
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Self-supervised learning (SSL) has shown remarkable performance in computer vision tasks when trained offline. However, in a Continual Learning (CL) scenario where new data is introduced progressively, models still suffer from catastrophic forgetting. Retraining a model from scratch to adapt to newly generated data is time-consuming and inefficient. Previous approaches suggested re-purposing self-supervised objectives with knowledge distillation to mitigate forgetting across tasks, assuming that labels from all tasks are available during fine-tuning. In this paper, we generalize self-supervised continual learning in a practical setting where available labels can be leveraged in any step of the SSL process. With an increasing number of continual tasks, this offers more flexibility in the pre-training and fine-tuning phases. With Kaizen(1), we introduce a training architecture that is able to mitigate catastrophic forgetting for both the feature extractor and classifier with a carefully designed loss function. By using a set of comprehensive evaluation metrics reflecting different aspects of continual learning, we demonstrated that Kaizen significantly outperforms previous SSL models in competitive vision benchmarks, with up to 16.5% accuracy improvement on split CIFAR-100. Kaizen is able to balance the trade-off between knowledge retention and learning from new data with an end-to-end model, paving the way for practical deployment of continual learning systems.
引用
收藏
页码:2829 / 2838
页数:10
相关论文
共 50 条
  • [1] SPeCiaL: Self-supervised Pretraining for Continual Learning
    Caccia, Lucas
    Pineau, Joelle
    CONTINUAL SEMI-SUPERVISED LEARNING, CSSL 2021, 2022, 13418 : 91 - 103
  • [2] Continual Barlow Twins: Continual Self-Supervised Learning for Remote Sensing Semantic Segmentation
    Marsocci, Valerio
    Scardapane, Simone
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2023, 16 : 5049 - 5060
  • [3] Self-Supervised Models are Continual Learners
    Fini, Enrico
    da Costa, Victor G. Turrisi
    Alameda-Pineda, Xavier
    Ricci, Elisa
    Alahari, Karteek
    Mairal, Julien
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 9611 - 9620
  • [4] Alleviating Representational Shift for Continual Fine-tuning
    Jie, Shibo
    Deng, Zhi-Hong
    Li, Ziheng
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2022, 2022, : 3809 - 3818
  • [5] CONTINUAL SELF-SUPERVISED LEARNING IN EARTH OBSERVATION WITH EMBEDDING REGULARIZATION
    Moieez, Hamna
    Marsocci, Valerio
    Scardapane, Simone
    IGARSS 2023 - 2023 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM, 2023, : 5029 - 5032
  • [6] Continual Robot Learning Using Self-Supervised Task Inference
    Hafez, Muhammad Burhan
    Wermter, Stefan
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2024, 16 (03) : 947 - 960
  • [7] Self-Supervised Continual Graph Learning in Adaptive Riemannian Spaces
    Sun, Li
    Ye, Junda
    Peng, Hao
    Wang, Feiyang
    Yu, Philip S.
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 4, 2023, : 4633 - 4642
  • [8] Learning Representations for New Sound Classes With Continual Self-Supervised Learning
    Wang, Zhepei
    Subakan, Cem
    Jiang, Xilin
    Wu, Junkai
    Tzinis, Efthymios
    Ravanelli, Mirco
    Smaragdis, Paris
    IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 2607 - 2611
  • [9] Improving Pedestrian Prediction Models With Self-Supervised Continual Learning
    Knoedler, Luzia
    Salmi, Chadi
    Zhu, Hai
    Brito, Bruno
    Alonso-Mora, Javier
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (02): : 4781 - 4788
  • [10] DLCFT: Deep Linear Continual Fine-Tuning for General Incremental Learning
    Shon, Hyounguk
    Lee, Janghyeon
    Kim, Seung Hwan
    Kim, Junmo
    COMPUTER VISION - ECCV 2022, PT XXXIII, 2022, 13693 : 513 - 529