Learning to Prompt for Continual Learning

被引:252
|
作者
Wang, Zifeng [1 ]
Zhang, Zizhao [2 ]
Lee, Hen Yu [2 ]
Zhang, Han [3 ]
Sun, Ruoxi [2 ]
Ren, Xiaoqi [2 ]
Su, Guolong [3 ]
Perot, Vincent [3 ]
Dy, Jennifer [1 ]
Pfister, Tomas [2 ]
机构
[1] Northeastern Univ, Boston, MA 02115 USA
[2] Google Cloud AI, Sunnyvale, CA USA
[3] Google Res, Sunnyvale, CA USA
关键词
SYSTEMS;
D O I
10.1109/CVPR52688.2022.00024
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The mainstream paradigm behind continual learning has been to adapt the model parameters to non-stationary data distributions, where catastrophic forgetting is the central challenge. Typical methods rely on a rehearsal buffer or known task identity at test time to retrieve learned knowledge and address forgetting, while this work presents a new paradigm for continual learning that aims to train a more succinct memory system without accessing task identity at test time. Our method learns to dynamically prompt (L2P) a pre-trained model to learn tasks sequentially under different task transitions. In our proposed framework, prompts are small learnable parameters, which are maintained in a memory space. The objective is to optimize prompts to instruct the model prediction and explicitly manage task-invariant and task-specific knowledge while maintaining model plasticity. We conduct comprehensive experiments under popular image classification benchmarks with different challenging continual learning settings, where L2P consistently outperforms prior state-ofthe-art methods. Surprisingly, L2P achieves competitive results against rehearsal-based methods even without a rehearsal buffer and is directly applicable to challenging taskagnostic continual learning. Source code is available at https:// github.com/ google- research/l2p.
引用
收藏
页码:139 / 149
页数:11
相关论文
共 50 条
  • [41] Memory Bounds for Continual Learning
    Chen, Xi
    Papadimitriou, Christos
    Peng, Binghui
    2022 IEEE 63RD ANNUAL SYMPOSIUM ON FOUNDATIONS OF COMPUTER SCIENCE (FOCS), 2022, : 519 - 530
  • [42] Continual Learning Based on Knowledge Distillation and Representation Learning
    Chen, Xiu-Yan
    Liu, Jian-Wei
    Li, Wen-Tao
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2022, PT IV, 2022, 13532 : 27 - 38
  • [43] Continual Learning Dialogue Systems - Learning during Conversation
    Mazumder, Sahisnu
    Liu, Bing
    PROCEEDINGS OF THE 45TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '22), 2022, : 3429 - 3432
  • [44] Prompt Distribution Learning
    Lu, Yuning
    Liu, Jianzhuang
    Zhang, Yonggang
    Liu, Yajing
    Tian, Xinmei
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 5196 - 5205
  • [45] Continual learning with selective nets
    Luu, Hai Tung
    Szemenyei, Marton
    APPLIED INTELLIGENCE, 2025, 55 (07)
  • [46] Continual Learning of Object Instances
    Parshotam, Kishan
    Kilickaya, Mert
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), 2020, : 907 - 914
  • [47] CONTINUAL LEARNING AND PRIVATE UNLEARNING
    Liu, Bo
    Liu, Qiang
    Stone, Peter
    CONFERENCE ON LIFELONG LEARNING AGENTS, VOL 199, 2022, 199
  • [48] Adversary Aware Continual Learning
    Umer, Muhammad
    Polikar, Robi
    IEEE ACCESS, 2024, 12 : 126108 - 126121
  • [49] Subspace distillation for continual learning
    Roy, Kaushik
    Simon, Christian
    Moghadam, Peyman
    Harandi, Mehrtash
    NEURAL NETWORKS, 2023, 167 : 65 - 79
  • [50] The Present and Future of Continual Learning
    Bae, Heechul
    Song, Soonyong
    Park, Junhee
    11TH INTERNATIONAL CONFERENCE ON ICT CONVERGENCE: DATA, NETWORK, AND AI IN THE AGE OF UNTACT (ICTC 2020), 2020, : 1193 - 1195