FedViT: Federated continual learning of vision transformer at edge

被引:2
|
作者
Zuo, Xiaojiang [1 ]
Luopan, Yaxin [1 ]
Han, Rui [1 ]
Zhang, Qinglong [1 ]
Liu, Chi Harold [1 ]
Wang, Guoren [1 ]
Chen, Lydia Y. [2 ]
机构
[1] Beijing Inst Technol, Beijing, Peoples R China
[2] Delft Univ Technol, Delft, Netherlands
基金
国家重点研发计划;
关键词
Catastrophic forgetting; Continual learning; Edge computing; Federated learning; Knowledge transfer negative; Vision transformer;
D O I
10.1016/j.future.2023.11.038
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Deep Neural Networks (DNNs) have been ubiquitously adopted in internet of things and are becoming an integral part of our daily life. When tackling the evolving learning tasks in real world, such as classifying different types of objects, DNNs face the challenge to continually retrain themselves according to the tasks on different edge devices. Federated continual learning (FCL) is a promising technique that offers partial solutions but yet to overcome the following difficulties: the significant accuracy loss due to the limited on -device processing, the negative knowledge transfer caused by the limited communication of non-IID (non -Independent and Identically Distributed) data, and the limited scalability on the tasks and edge devices. Moreover, existing FCL techniques are designed for convolutional neural networks (CNNs), which have not utilized the full potential of newly emerged powerful vision transformers (ViTs). Considering ViTs depend heavily on training data diversity and volume, we hypothesize ViTs are well -suited for FCL where data arrives continually. In this paper, we propose FedViT, an accurate and scalable federated continual learning framework for ViT models, via a novel concept of signature task knowledge. FedViT is a client -side solution that continuously extracts and integrates the knowledge of signature tasks which are highly influenced by the current task. Each client of FedViT is composed of a knowledge extractor, a gradient restorer and, most importantly, a gradient integrator. Upon training for a new task, the gradient integrator ensures the prevention of catastrophic forgetting and mitigation of negative knowledge transfer by effectively combining signature tasks identified from the past local tasks and other clients' current tasks through the global model. We implement FedViT in PyTorch and extensively evaluate it against state-of-the-art techniques using popular federated continual learning benchmarks. Extensive evaluation results on heterogeneous edge devices show that FedViT improves model accuracy by 88.61% without increasing model training time, reduces communication cost by 61.55%, and achieves more improvements under difficult scenarios such as large numbers of tasks or clients, and training different complex ViT models.
引用
收藏
页码:1 / 15
页数:15
相关论文
共 50 条
  • [1] CONTINUAL LEARNING IN VISION TRANSFORMER
    Takeda, Mana
    Yanai, Keiji
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 616 - 620
  • [2] Continual Learning with Lifelong Vision Transformer
    Wang, Zhen
    Liu, Liu
    Duan, Yiqun
    Kong, Yajing
    Tao, Dacheng
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 171 - 181
  • [3] Online Continual Learning with Contrastive Vision Transformer
    Wang, Zhen
    Liu, Liu
    Kong, Yajing
    Guo, Jiaxian
    Tao, Dacheng
    [J]. COMPUTER VISION, ECCV 2022, PT XX, 2022, 13680 : 631 - 650
  • [4] Energy-Efficient Personalized Federated Continual Learning on Edge
    Yang, Zhao
    Wang, Haoyang
    Sun, Qingshuang
    [J]. IEEE Embedded Systems Letters, 2024, 16 (04) : 345 - 348
  • [5] FeSViBS: Federated Split Learning of Vision Transformer with Block Sampling
    Almalik, Faris
    Alkhunaizi, Naif
    Almakky, Ibrahim
    Nandakumar, Karthik
    [J]. MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2023, PT II, 2023, 14221 : 350 - 360
  • [6] Task-Free Dynamic Sparse Vision Transformer for Continual Learning
    Ye, Fei
    Bors, Adrian G.
    [J]. THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 15, 2024, : 16442 - 16450
  • [7] POSTER: Advancing Federated Edge Computing with Continual Learning for Secure and Efficient Performance
    Chen, Chunlu
    Wang, Kevin I-Kai
    Li, Peng
    Sakurai, Kouichi
    [J]. APPLIED CRYPTOGRAPHY AND NETWORK SECURITY WORKSHOPS, ACNS 2023 SATELLITE WORKSHOPS, ADSC 2023, AIBLOCK 2023, AIHWS 2023, AIOTS 2023, CIMSS 2023, CLOUD S&P 2023, SCI 2023, SECMT 2023, SIMLA 2023, 2023, 13907 : 685 - 689
  • [8] Federated probability memory recall for federated continual learning
    Wang, Zhe
    Zhang, Yu
    Xu, Xinlei
    Fu, Zhiling
    Yang, Hai
    Du, Wenli
    [J]. INFORMATION SCIENCES, 2023, 629 : 551 - 565
  • [9] Federated Continual Learning for Socially Aware Robotics
    Guerdan, Luke
    Gunes, Hatice
    [J]. 2023 32ND IEEE INTERNATIONAL CONFERENCE ON ROBOT AND HUMAN INTERACTIVE COMMUNICATION, RO-MAN, 2023, : 1522 - 1529
  • [10] Ensemble and continual federated learning for classification tasks
    Fernando E. Casado
    Dylan Lema
    Roberto Iglesias
    Carlos V. Regueiro
    Senén Barro
    [J]. Machine Learning, 2023, 112 : 3413 - 3453