S-Prompts Learning with Pre-trained Transformers: An Occam's Razor for Domain Incremental Learning

被引:0
|
作者
Wang, Yabin [1 ,2 ]
Huang, Zhiwu [2 ]
Hong, Xiaopeng [1 ,3 ,4 ]
机构
[1] Xi An Jiao Tong Univ, Xian, Peoples R China
[2] Singapore Management Univ, Singapore, Singapore
[3] Harbin Inst Technol, Harbin, Peoples R China
[4] Pengcheng Lab, Shenzhen, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
State-of-the-art deep neural networks are still struggling to address the catastrophic forgetting problem in continual learning. In this paper, we propose one simple paradigm (named as S-Prompting) and two concrete approaches to highly reduce the forgetting degree in one of the most typical continual learning scenarios, i.e., domain increment learning (DIL). The key idea of the paradigm is to learn prompts independently across domains with pre-trained transformers, avoiding the use of exemplars that commonly appear in conventional methods. This results in a win-win game where the prompting can achieve the best for each domain. The independent prompting across domains only requests one single cross-entropy loss for training and one simple K-NN operation as a domain identifier for inference. The learning paradigm derives an image prompt learning approach and a novel language-image prompt learning approach. Owning an excellent scalability (0.03% parameter increase per domain), the best of our approaches achieves a remarkable relative improvement (an average of about 30%) over the best of the state-of-the-art exemplar-free methods for three standard DIL tasks, and even surpasses the best of them relatively by about 6% in average when they use exemplars. Source code is available at https://github.com/iamwangyabin/S-Prompts.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Occam's Razor in sensorimotor learning
    Genewein, Tim
    Braun, Daniel A.
    PROCEEDINGS OF THE ROYAL SOCIETY B-BIOLOGICAL SCIENCES, 2014, 281 (1783)
  • [2] Causal Learning with Occam's Razor
    Schulte, Oliver
    STUDIA LOGICA, 2019, 107 (05) : 991 - 1023
  • [3] Causal Learning with Occam’s Razor
    Oliver Schulte
    Studia Logica, 2019, 107 : 991 - 1023
  • [4] Statistical Schema Learning using Occam's Razor
    Talbot, Justin
    Ting, Daniel
    PROCEEDINGS OF THE 2022 INTERNATIONAL CONFERENCE ON MANAGEMENT OF DATA (SIGMOD '22), 2022, : 176 - 189
  • [5] Class-Incremental Learning with Strong Pre-trained Models
    Wu, Tz-Ying
    Swaminathan, Gurumurthy
    Li, Zhizhong
    Ravichandran, Avinash
    Vasconcelos, Nuno
    Bhotika, Rahul
    Soatto, Stefano
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 9591 - 9600
  • [6] Word Representation Learning in Multimodal Pre-Trained Transformers: An Intrinsic Evaluation
    Pezzelle, Sandro
    Takmaz, Ece
    Fernandez, Raquel
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2021, 9 : 1563 - 1579
  • [7] Optimally trained regression trees and Occam's razor
    Savicky, P
    Klaschka, J
    COMPSTAT 2002: PROCEEDINGS IN COMPUTATIONAL STATISTICS, 2002, : 479 - 484
  • [8] Fusing Pre-trained Language Models with Multimodal Prompts through Reinforcement Learning
    Yu, Youngjae
    Chung, Jiwan
    Yun, Heeseung
    Hessel, Jack
    Park, Jae Sung
    Lu, Ximing
    Zellers, Rowan
    Ammanabrolu, Prithviraj
    Le Bras, Ronan
    Kim, Gunhee
    Choi, Yejin
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 10845 - 10856
  • [9] PAC Learning and Occam's Razor: Probably Approximately Incorrect
    Herrmann, Daniel A.
    PHILOSOPHY OF SCIENCE, 2020, 87 (04) : 685 - 703
  • [10] Investor's ESG tendency probed by pre-trained transformers
    Li, Chao
    Keeley, Alexander Ryota
    Takeda, Shutaro
    Seki, Daikichi
    Managi, Shunsuke
    CORPORATE SOCIAL RESPONSIBILITY AND ENVIRONMENTAL MANAGEMENT, 2025, 32 (02) : 2051 - 2071