Learning to Prompt for Continual Learning

被引:252
|
作者
Wang, Zifeng [1 ]
Zhang, Zizhao [2 ]
Lee, Hen Yu [2 ]
Zhang, Han [3 ]
Sun, Ruoxi [2 ]
Ren, Xiaoqi [2 ]
Su, Guolong [3 ]
Perot, Vincent [3 ]
Dy, Jennifer [1 ]
Pfister, Tomas [2 ]
机构
[1] Northeastern Univ, Boston, MA 02115 USA
[2] Google Cloud AI, Sunnyvale, CA USA
[3] Google Res, Sunnyvale, CA USA
关键词
SYSTEMS;
D O I
10.1109/CVPR52688.2022.00024
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The mainstream paradigm behind continual learning has been to adapt the model parameters to non-stationary data distributions, where catastrophic forgetting is the central challenge. Typical methods rely on a rehearsal buffer or known task identity at test time to retrieve learned knowledge and address forgetting, while this work presents a new paradigm for continual learning that aims to train a more succinct memory system without accessing task identity at test time. Our method learns to dynamically prompt (L2P) a pre-trained model to learn tasks sequentially under different task transitions. In our proposed framework, prompts are small learnable parameters, which are maintained in a memory space. The objective is to optimize prompts to instruct the model prediction and explicitly manage task-invariant and task-specific knowledge while maintaining model plasticity. We conduct comprehensive experiments under popular image classification benchmarks with different challenging continual learning settings, where L2P consistently outperforms prior state-ofthe-art methods. Surprisingly, L2P achieves competitive results against rehearsal-based methods even without a rehearsal buffer and is directly applicable to challenging taskagnostic continual learning. Source code is available at https:// github.com/ google- research/l2p.
引用
收藏
页码:139 / 149
页数:11
相关论文
共 50 条
  • [11] Dual Prompt Learning for Continual Rain Removal from Single Images
    Liu, Minghao
    Yang, Wenhan
    Hu, Yuzhang
    Liu, Jiaying
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 7215 - +
  • [12] Towards prompt tuning-based software vulnerability assessment with continual learning
    Xue, Jiacheng
    Chen, Xiang
    Wang, Jiyu
    Cui, Zhanqi
    COMPUTERS & SECURITY, 2025, 150
  • [13] CONTINUAL LEARNING
    BROWN, WE
    JOURNAL OF THE AMERICAN DENTAL ASSOCIATION, 1965, 71 (04): : 935 - &
  • [14] Continual learning
    King, Denise
    JOURNAL OF EMERGENCY NURSING, 2008, 34 (04) : 283 - 283
  • [15] CODA-Prompt: COntinual Decomposed Attention-based Prompting for Rehearsal-Free Continual Learning
    Smith, James Seale
    Karlinsky, Leonid
    Gutta, Vyshnavi
    Cascante-Bonilla, Paola
    Kim, Donghyun
    Arbelle, Assaf
    Panda, Rameswar
    Feris, Rogerio
    Kira, Zsolt
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 11909 - 11919
  • [16] TASK-WISE PROMPT QUERY FUNCTION FOR REHEARSAL-FREE CONTINUAL LEARNING
    Chen, Shuai
    Zhang, Mingyi
    Zhang, Junge
    Huang, Kaiqi
    2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, ICASSP 2024, 2024, : 6320 - 6324
  • [17] Continual Few-Shot Relation Extraction with Prompt-Based Contrastive Learning
    Wu, Fei
    Zhang, Chong
    Tan, Zhen
    Xu, Hao
    Ge, Bin
    WEB AND BIG DATA, PT IV, APWEB-WAIM 2023, 2024, 14334 : 312 - 327
  • [18] Progressive learning: A deep learning framework for continual learning
    Fayek, Haytham M.
    Cavedon, Lawrence
    Wu, Hong Ren
    NEURAL NETWORKS, 2020, 128 : 345 - 357
  • [19] Logarithmic Continual Learning
    Masarczyk, Wojciech
    Wawrzynski, Pawel
    Marczak, Daniel
    Deja, Kamil
    Trzcinski, Tomasz
    IEEE ACCESS, 2022, 10 : 117001 - 117010
  • [20] Bilevel Continual Learning
    Shaker, Ammar
    Alesiani, Francesco
    Yu, Shujian
    Yin, Wenzhe
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,