Recent Advances of Foundation Language Models-based Continual Learning: A Survey

被引:0
|
作者
Yang, Yutao [1 ]
Zhou, Jie [1 ]
Ding, Xuan wen [1 ]
Huai, Tianyu [1 ]
Liu, Shunyu [1 ]
Chen, Qin [1 ]
Xie, Yuan [1 ]
He, Liang [1 ]
机构
[1] East China Normal Univ, Sch Comp Sci & Technol, Shanghai, Peoples R China
关键词
Continual learning; foundation language models; pre-trained language models; large language models; vision-language models; survey; NEURAL-NETWORKS; LIFELONG;
D O I
10.1145/3705725
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Recently, foundation language models (LMs) have marked significant achievements in the domains of natural language processing and computer vision. Unlike traditional neural network models, foundation LMs obtain a great ability for transfer learning by acquiring rich common sense knowledge through pre-training on extensive unsupervised datasets with a vast number of parameters. Despite these capabilities, LMs still struggle with catastrophic forgetting, hindering their ability to learn continuously like humans. To address this, continual learning (CL) methodologies have been introduced, allowing LMs to adapt to new tasks while retaining learned knowledge. However, a systematic taxonomy of existing approaches and a comparison of their performance are still lacking. In this article, we delve into a comprehensive review, summarization, and classification of the existing literature on CL-based approaches applied to foundation language models, such as pre-trained language models, large language models, and vision-language models. We divide these studies into offline and online CL, which consist of traditional methods, parameter-efficient-based methods, instruction tuning-based methods and continual pre-training methods. Additionally, we outline the typical datasets and metrics employed in CL research and provide a detailed analysis of the challenges and future work for LMs-based continual learning.
引用
收藏
页数:38
相关论文
共 50 条
  • [1] Recent advances in deep learning and language models for studying the microbiome
    Yan, Binghao
    Nam, Yunbi
    Li, Lingyao
    Deek, Rebecca A.
    Li, Hongzhe
    Ma, Siyuan
    FRONTIERS IN GENETICS, 2025, 15
  • [2] Geospatial Foundation Models: Recent Advances and Applications
    Vatsavai, Ranga Raju
    PROCEEDINGS OF THE 12TH ACM SIGSPATIAL INTERNATIONAL WORKSHOP ON ANALYTICS FOR BIG GEOSPATIAL DATA, BIGSPATIAL 2024, 2024, : 30 - 33
  • [3] An advanced deep learning models-based plant disease detection: A review of recent research
    Shoaib, Muhammad
    Shah, Babar
    EI-Sappagh, Shaker
    Ali, Akhtar
    Ullah, Asad
    Alenezi, Fayadh
    Gechev, Tsanko
    Hussain, Tariq
    Ali, Farman
    FRONTIERS IN PLANT SCIENCE, 2023, 14
  • [4] CONTINUAL LEARNING WITH FOUNDATION MODELS: AN EMPIRICAL STUDY OF LATENT REPLAY
    Ostapenko, Oleksiy
    Lesort, Timothee
    Rodriguez, Pau
    Arefin, Md Rifat
    Douillard, Arthur
    Rish, Irina
    Charlin, Laurent
    CONFERENCE ON LIFELONG LEARNING AGENTS, VOL 199, 2022, 199
  • [5] Continual Learning with Pre-Trained Models: A Survey
    Zhou, Da-Wei
    Sun, Hai-Long
    Ning, Jingyi
    Ye, Han-Jia
    Zhan, De-Chuan
    PROCEEDINGS OF THE THIRTY-THIRD INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2024, 2024, : 8363 - 8371
  • [6] A survey on recent advances in Sign Language Production
    Rastgoo, Razieh
    Kiani, Kourosh
    Escalera, Sergio
    Athitsos, Vassilis
    Sabokrou, Mohammad
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 243
  • [7] Recent Advances in Natural Language Processing via Large Pre-trained Language Models: A Survey
    Min, Bonan
    Ross, Hayley
    Sulem, Elior
    Ben Veyseh, Amir Pouran
    Nguyen, Thien Huu
    Sainz, Oscar
    Agirre, Eneko
    Heintz, Ilana
    Roth, Dan
    ACM COMPUTING SURVEYS, 2024, 56 (02)
  • [8] Optimizing Foundation Models for Histopathology: A Continual Learning Approach to Cancer Detection
    Yaday, Ankur
    Daescu, Ovidiu
    TRUSTWORTHY ARTIFICIAL INTELLIGENCE FOR HEALTHCARE, TAI4H 2024, 2024, 14812 : 145 - 156
  • [9] Large Language Models-Based Local Explanations of Text Classifiers
    Angiulli, Fabrizio
    De Luca, Francesco
    Fassetti, Fabio
    Nistico, Simona
    DISCOVERY SCIENCE, DS 2024, PT I, 2025, 15243 : 19 - 35
  • [10] Deep learning models-based classification of solid waste
    Muthukrishnan, Anuradham M.
    Krishna, B.V. Santhosh
    Atmakuri, Murali Krishna
    Usha, D.
    International Journal of Internet Protocol Technology, 2024, 17 (01) : 19 - 30