FedKC: Federated Knowledge Composition for Multilingual Natural Language Understanding

被引:9
|
作者
Wang, Haoyu [1 ]
Zhao, Handong [2 ]
Wang, Yaqing [1 ]
Yu, Tong [2 ]
Gu, Jiuxiang [2 ]
Gao, Jing [1 ]
机构
[1] Purdue Univ, W Lafayette, IN 47907 USA
[2] Adobe Res, San Jose, CA USA
基金
美国国家科学基金会;
关键词
Federated learning; Multilingual natural language understanding;
D O I
10.1145/3485447.3511988
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Multilingual natural language understanding, which aims to comprehend multilingual documents, is an important task. Existing efforts have been focusing on the analysis of centrally stored text data, but in real practice, multilingual data is usually distributed. Federated learning is a promising paradigm to solve this problem, which trains local models with decentralized data on local clients and aggregates local models on the central server to achieve a good global model. However, existing federated learning methods assume that data are independent and identically distributed (IID), and cannot handle multilingual data, that are usually non-IID with severely skewed distributions: First, multilingual data is stored on local client devices such that there are only monolingual or bilingual data stored on each client. This makes it difficult for local models to know the information of documents in other languages. Second, the distribution over different languages could be skewed. High resource language data is much more abundant than low resource language data. The model trained on such skewed data may focus more on high resource languages but fail to consider the key information of low resource languages. To solve the aforementioned challenges of multilingual federated NLU, we propose a plug-and-play knowledge composition (KC) module, called FedKC, which exchanges knowledge among clients without sharing raw data. Specifically, we propose an effective way to calculate a consistency loss defined based on the shared knowledge across clients, which enables models trained on different clients achieve similar predictions on similar data. Leveraging this consistency loss, joint training is thus conducted on distributed data respecting the privacy constraints. We also analyze the potential risk of FedKC and provide theoretical bound to show that it is difficult to recover data from the corrupted data. We conduct extensive experiments on three public multilingual datasets for three typical NLU tasks, including paraphrase identification, question answering matching, and news classification. The experiment results show that the proposed FedKC can outperform state-of-the-art baselines on the three datasets significantly.
引用
收藏
页码:1839 / 1850
页数:12
相关论文
共 50 条
  • [1] KNOWLEDGE REPRESENTATION FOR NATURAL LANGUAGE UNDERSTANDING
    Stanojevic, Mladen
    Vranes, Sanja
    FACTA UNIVERSITATIS-SERIES MATHEMATICS AND INFORMATICS, 2006, 21 : 93 - 104
  • [2] Cross-language relations in composition: Understanding the multilingual nature of composition courses
    Valfredini, Alessia
    JOURNAL OF SECOND LANGUAGE WRITING, 2012, 21 (01) : 87 - 89
  • [3] Collaborative multilingual knowledge management based on controlled natural language
    Kaljurand, Kaarel
    Kuhn, Tobias
    Canedo, Laura
    SEMANTIC WEB, 2015, 6 (03) : 241 - 258
  • [4] Extrapolating Multilingual Language Understanding Models as Multilingual Language Generators
    Wu, Bohong
    Yuan, Fei
    Zhao, Hai
    Li, Lei
    Xu, Jingjing
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 15432 - 15444
  • [5] Multilingual Natural Language Understanding for the FPT.AI Conversational Platform
    Le-Hong, Phuong
    Nguyen, Thi Thuy Lien
    Pham, Minh Tu
    Vu, Thanh Hai
    Proceedings - International Conference on Knowledge and Systems Engineering, KSE, 2022, 2022-October
  • [6] Knowledge-based Data Processing for Multilingual Natural Language Analysis
    Jain, Deepak Kumar
    Eyre, Yamila Garcia-Martinez
    Kumar, Akshi
    Gupta, Brij B.
    Kotecha, Ketan
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2024, 23 (05)
  • [8] Recent Trends in Natural Language Understanding for Procedural Knowledge
    Mujtaba, Dena F.
    Mahapatra, Nihar R.
    2019 6TH INTERNATIONAL CONFERENCE ON COMPUTATIONAL SCIENCE AND COMPUTATIONAL INTELLIGENCE (CSCI 2019), 2019, : 420 - 424
  • [9] Language model for multilingual natural language generation
    Zhang, Dongmo
    Ge, Yong
    Yao, Tianfang
    Shanghai Jiaotong Daxue Xuebao/Journal of Shanghai Jiaotong University, 2000, 34 (07): : 944 - 947
  • [10] Multilingual LAMA: Investigating Knowledge in Multilingual Pretrained Language Models
    Kassner, Nora
    Dufter, Philipp
    Schutze, Hinrich
    16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 3250 - 3258