Cross-lingual transfer learning: A PARAFAC2 approach

被引:1
|
作者
Pantraki, Evangelia [1 ]
Tsingalis, Ioannis [1 ]
Kotropoulos, Constantine [1 ]
机构
[1] Aristotle Univ Thessaloniki, Dept Informat, Thessaloniki, Greece
关键词
PARAFAC2; Cross-lingual transfer learning; Cross-lingual document classification; Cross-lingual authorship attribution; Language processing; EMBEDDINGS;
D O I
10.1016/j.patrec.2022.05.008
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The proposed framework addresses the problem of cross-lingual transfer learning resorting to Parallel Factor Analysis 2 (PARAFAC2). To avoid the need for multilingual parallel corpora, a pairwise setting is adopted where a PARAFAC2 model is fitted to documents written in English (source language) and a different target language. Firstly, an unsupervised PARAFAC2 model is fitted to parallel unlabelled corpora pairs to learn the latent relationship between the source and target language. The fitted model is used to create embeddings for a text classification task (document classification or authorship attribution). Subsequently, a logistic regression classifier is fitted to the training source language embeddings and tested on the training target language embeddings. Following the zero-shot setting, no labels are exploited for the target language documents. The proposed framework incorporates a self-learning process by utilizing the predicted labels as pseudo-labels to train a new, pseudo-supervised PARAFAC2 model, which aims to extract latent class-specific information while fusing language-specific information. Thorough evaluation is conducted on cross-lingual document classification and cross-lingual authorship attribution. Remarkably, the proposed framework achieves competitive results when compared to deep learning methods in cross-lingual transfer learning tasks. (C) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页码:167 / 173
页数:7
相关论文
共 50 条
  • [31] CROSS-LINGUAL TRANSFER LEARNING FOR LOW-RESOURCE SPEECH TRANSLATION
    Khurana, Sameer
    Dawalatabad, Nauman
    Laurent, Antoine
    Vicente, Luis
    Gimeno, Pablo
    Mingote, Victoria
    Glass, James
    2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING WORKSHOPS, ICASSPW 2024, 2024, : 670 - 674
  • [32] An Approach to Cross-Lingual Voice Conversion
    Rallabandi, Sai Sirisha
    Gangashetty, Suryakanth V.
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [33] A Learning to rank framework based on cross-lingual loss function for cross-lingual information retrieval
    Ghanbari, Elham
    Shakery, Azadeh
    APPLIED INTELLIGENCE, 2022, 52 (03) : 3156 - 3174
  • [34] A Deep Transfer Learning Method for Cross-Lingual Natural Language Inference
    Bandyopadhyay, Dibyanayan
    De, Arkadipta
    Gain, Baban
    Saikh, Tanik
    Ekbal, Asif
    LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 3084 - 3092
  • [35] PARAFAC2 - Part I. A direct fitting algorithm for the PARAFAC2 model
    Kiers, HAL
    Ten Berge, JMF
    Bro, R
    JOURNAL OF CHEMOMETRICS, 1999, 13 (3-4) : 275 - 294
  • [36] Building Indonesian Dependency Parser Using Cross-lingual Transfer Learning
    Maulana, Andhika Yusup
    Alfina, Ika
    Azizah, Kurniawati
    2022 INTERNATIONAL CONFERENCE ON ASIAN LANGUAGE PROCESSING (IALP 2022), 2022, : 488 - 493
  • [37] Cross-lingual transfer learning for relation extraction using Universal Dependencies
    Taghizadeh, Nasrin
    Faili, Heshaam
    COMPUTER SPEECH AND LANGUAGE, 2022, 71
  • [38] DiTTO : A Feature Representation Imitation Approach for Improving Cross-Lingual Transfer
    Kumar, Shanu
    Soujanya, Abbaraju
    Dandapat, Sandipan
    Sitaram, Sunayana
    Choudhury, Monojit
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 385 - 406
  • [39] A Learning to rank framework based on cross-lingual loss function for cross-lingual information retrieval
    Elham Ghanbari
    Azadeh Shakery
    Applied Intelligence, 2022, 52 : 3156 - 3174
  • [40] A Multi-media Approach to Cross-lingual Entity Knowledge Transfer
    Lu, Di
    Pan, Xiaoman
    Pourdamghani, Nima
    Chang, Shih-Fu
    Ji, Heng
    Knight, Kevin
    PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, 2016, : 54 - 65