FedKC: Federated Knowledge Composition for Multilingual Natural Language Understanding

被引:9
|
作者
Wang, Haoyu [1 ]
Zhao, Handong [2 ]
Wang, Yaqing [1 ]
Yu, Tong [2 ]
Gu, Jiuxiang [2 ]
Gao, Jing [1 ]
机构
[1] Purdue Univ, W Lafayette, IN 47907 USA
[2] Adobe Res, San Jose, CA USA
基金
美国国家科学基金会;
关键词
Federated learning; Multilingual natural language understanding;
D O I
10.1145/3485447.3511988
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Multilingual natural language understanding, which aims to comprehend multilingual documents, is an important task. Existing efforts have been focusing on the analysis of centrally stored text data, but in real practice, multilingual data is usually distributed. Federated learning is a promising paradigm to solve this problem, which trains local models with decentralized data on local clients and aggregates local models on the central server to achieve a good global model. However, existing federated learning methods assume that data are independent and identically distributed (IID), and cannot handle multilingual data, that are usually non-IID with severely skewed distributions: First, multilingual data is stored on local client devices such that there are only monolingual or bilingual data stored on each client. This makes it difficult for local models to know the information of documents in other languages. Second, the distribution over different languages could be skewed. High resource language data is much more abundant than low resource language data. The model trained on such skewed data may focus more on high resource languages but fail to consider the key information of low resource languages. To solve the aforementioned challenges of multilingual federated NLU, we propose a plug-and-play knowledge composition (KC) module, called FedKC, which exchanges knowledge among clients without sharing raw data. Specifically, we propose an effective way to calculate a consistency loss defined based on the shared knowledge across clients, which enables models trained on different clients achieve similar predictions on similar data. Leveraging this consistency loss, joint training is thus conducted on distributed data respecting the privacy constraints. We also analyze the potential risk of FedKC and provide theoretical bound to show that it is difficult to recover data from the corrupted data. We conduct extensive experiments on three public multilingual datasets for three typical NLU tasks, including paraphrase identification, question answering matching, and news classification. The experiment results show that the proposed FedKC can outperform state-of-the-art baselines on the three datasets significantly.
引用
收藏
页码:1839 / 1850
页数:12
相关论文
共 50 条
  • [41] XINFOTABS: Evaluating Multilingual Tabular Natural Language Inference
    Minhas, Bhavnick
    Shankhdhar, Anant
    Gupta, Vivek
    Aggrawal, Divyanshu
    Zhang, Shuo
    PROCEEDINGS OF THE FIFTH FACT EXTRACTION AND VERIFICATION WORKSHOP (FEVER 2022), 2022, : 59 - 77
  • [42] DKPLM: Decomposable Knowledge-Enhanced Pre-trained Language Model for Natural Language Understanding
    Zhang, Taolin
    Wang, Chengyu
    Hu, Nan
    Qiu, Minghui
    Tang, Chengguang
    He, Xiaofeng
    Huang, Jun
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 11703 - 11711
  • [43] A Train-on-Target Strategy for Multilingual Spoken Language Understanding
    Garcia-Granada, Fernando
    Segarra, Encarna
    Millan, Carlos
    Sanchis, Emilio
    Hurtado, Lluis-F.
    ADVANCES IN SPEECH AND LANGUAGE TECHNOLOGIES FOR IBERIAN LANGUAGES, IBERSPEECH 2016, 2016, 10077 : 224 - 233
  • [44] Critical multilingual language awareness: the role of teachers as language activists and knowledge generators
    Cummins, Jim
    LANGUAGE AWARENESS, 2023, 32 (04) : 560 - 573
  • [45] Multilingual Knowledge Graph Completion from Pretrained Language Models with Knowledge Constraints
    Song, Ran
    He, Shizhu
    Gao, Shengxiang
    Cai, Li
    Liu, Kang
    Yu, Zhengtao
    Zhao, Jun
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 7709 - 7721
  • [46] NATURAL-LANGUAGE UNDERSTANDING
    WALDROP, MM
    SCIENCE, 1984, 224 (4647) : 372 - 374
  • [47] UNDERSTANDING NATURAL-LANGUAGE
    HAUGELAND, J
    JOURNAL OF PHILOSOPHY, 1979, 76 (11): : 619 - 632
  • [48] UNDERSTANDING NATURAL-LANGUAGE
    PITRAT, J
    RECHERCHE, 1978, 9 (93): : 876 - 881
  • [49] UNDERSTANDING OF NATURAL LANGUAGE BY COMPUTERS
    BERKELEY, EC
    COMPUTERS AND AUTOMATION, 1973, 22 (11): : 6 - 6
  • [50] Personalized Natural Language Understanding
    Liu, Xiaohu
    Sarikaya, Ruhi
    Zhao, Liang
    Ni, Yong
    Pan, Yi-Cheng
    17TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2016), VOLS 1-5: UNDERSTANDING SPEECH PROCESSING IN HUMANS AND MACHINES, 2016, : 1146 - 1150