FEDBERT: When Federated Learning Meets Pre-training

被引:47
|
作者
Tian, Yuanyishu [1 ]
Wan, Yao [1 ]
Lyu, Lingjuan [2 ]
Yao, Dezhong [1 ]
Jin, Hai [1 ]
Sun, Lichao [3 ]
机构
[1] Huazhong Univ Sci & Technol, Serv Comp Technol & Syst Lab, Natl Engn Res Ctr Big Data Technol & Syst, Sch Comp Sci & Technol,Cluster & Grid Comp Lab, 1037 Luoyu Rd, Wuhan 430074, Peoples R China
[2] Sony AI, Minato Ku, 1-7-1 Konan, Tokyo, Japan
[3] Lehigh Univ, 113 Res Dr, Bethlehem, PA 18015 USA
基金
中国国家自然科学基金;
关键词
Federated learning; pre-training; BERT; NLP;
D O I
10.1145/3510033
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The fast growth of pre-trained models (PTMs) has brought natural language processing to a new era, which has become a dominant technique for various natural language processing (NLP) applications. Every user can download the weights of PTMs, then fine-tune the weights for a task on the local side. However, the pre-training of a model relies heavily on accessing a large-scale of training data and requires a vast amount of computing resources. These strict requirements make it impossible for any single client to pre-train such a model. To grant clients with limited computing capability to participate in pre-training a large model, we propose a new learning approach, FEDBERT, that takes advantage of the federated learning and split learning approaches, resorting to pre-training BERT in a federated way. FEDBERT can prevent sharing the raw data information and obtain excellent performance. Extensive experiments on seven GLUE tasks demonstrate that FEDBERT can maintain its effectiveness without communicating to the sensitive local data of clients.
引用
收藏
页数:26
相关论文
共 50 条
  • [21] Integer Is Enough: When Vertical Federated Learning Meets Rounding
    Qiu, Pengyu
    Pu, Yuwen
    Liu, Yongchao
    Liu, Wenyan
    Yue, Yun
    Zhu, Xiaowei
    Li, Lichun
    Li, Jinbao
    Ji, Shouling
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 13, 2024, : 14704 - 14712
  • [22] When Federated Learning Meets Privacy-Preserving Computation
    Chen, Jingxue
    Yan, Hang
    Liu, Zhiyuan
    Zhang, Min
    Xiong, Hu
    Yu, Shui
    ACM COMPUTING SURVEYS, 2024, 56 (12)
  • [23] Belt and Braces: When Federated Learning Meets Differential Privacy
    Ren, Xuebin
    Yang, Shusen
    Zhao, Cong
    Mccann, Julie
    Xu, Zongben
    COMMUNICATIONS OF THE ACM, 2024, 67 (12) : 66 - 77
  • [24] When Federated Learning Meets Vision: An Outlook on Opportunities and Challenges
    Khan, Ahsan Raza
    Zoha, Ahmed
    Mohjazi, Lina
    Sajid, Hasan
    Abbasi, Qammar
    Imran, Muhammad Ali
    BODY AREA NETWORKS: SMART IOT AND BIG DATA FOR INTELLIGENT HEALTH MANAGEMENT, 2022, 420 : 308 - 319
  • [25] SLIP: Self-supervision Meets Language-Image Pre-training
    Mu, Norman
    Kirillov, Alexander
    Wagner, David
    Xie, Saining
    COMPUTER VISION, ECCV 2022, PT XXVI, 2022, 13686 : 529 - 544
  • [26] New Intent Discovery with Pre-training and Contrastive Learning
    Zhang, Yuwei
    Zhang, Haode
    Zhan, Li-Ming
    Wu, Xiao-Ming
    Lam, Albert Y. S.
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 256 - 269
  • [27] An Empirical Investigation of the Role of Pre-training in Lifelong Learning
    Mehta, Sanket Vaibhav
    Patil, Darshan
    Chandar, Sarath
    Strubell, Emma
    JOURNAL OF MACHINE LEARNING RESEARCH, 2023, 24
  • [28] Image Difference Captioning with Pre-training and Contrastive Learning
    Yao, Linli
    Wang, Weiying
    Jin, Qin
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 3108 - 3116
  • [29] Pre-training with Meta Learning for Chinese Word Segmentation
    Ke, Zhen
    Shi, Liang
    Sun, Songtao
    Meng, Erli
    Wang, Bin
    Qiu, Xipeng
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 5514 - 5523
  • [30] Pre-training Strategies and Datasets for Facial Representation Learning
    Bulat, Adrian
    Cheng, Shiyang
    Yang, Jing
    Garbett, Andrew
    Sanchez, Enrique
    Tzimiropoulos, Georgios
    COMPUTER VISION, ECCV 2022, PT XIII, 2022, 13673 : 107 - 125