Transformer Based Multi-Source Domain Adaptation

被引:0
|
作者
Wright, Dustin [1 ]
Augenstein, Isabelle [1 ]
机构
[1] Univ Copenhagen, Dept Comp Sci, Copenhagen, Denmark
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In practical machine learning settings, the data on which a model must make predictions often come from a different distribution than the data it was trained on. Here, we investigate the problem of unsupervised multi-source domain adaptation, where a model is trained on labelled data from multiple source domains and must make predictions on a domain for which no labelled data has been seen. Prior work with CNNs and RNNs has demonstrated the benefit of mixture of experts, where the predictions of multiple domain expert classifiers are combined; as well as domain adversarial training, to induce a domain agnostic representation space. Inspired by this, we investigate how such methods can be effectively applied to large pretrained transformer models. We find that domain adversarial training has an effect on the learned representations of these models while having little effect on their performance, suggesting that large transformer-based models are already relatively robust across domains. Additionally, we show that mixture of experts leads to significant performance improvements by comparing several variants of mixing functions, including one novel mixture based on attention. Finally, we demonstrate that the predictions of large pretrained transformer based domain experts are highly homogenous, making it challenging to learn effective functions for mixing their predictions.
引用
收藏
页码:7963 / 7974
页数:12
相关论文
共 50 条
  • [21] Graphical Modeling for Multi-Source Domain Adaptation
    Xu, Minghao
    Wang, Hang
    Ni, Bingbing
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (03) : 1727 - 1741
  • [22] Multi-Source Attention for Unsupervised Domain Adaptation
    Cui, Xia
    Bollegala, Danushka
    1ST CONFERENCE OF THE ASIA-PACIFIC CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 10TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (AACL-IJCNLP 2020), 2020, : 873 - 883
  • [23] Multi-Source Domain Adaptation with Mixture of Experts
    Guo, Jiang
    Shah, Darsh J.
    Barzilay, Regina
    2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 4694 - 4703
  • [24] Multi-source Domain Adaptation for Face Recognition
    Yi, Haiyang
    Xu, Zhi
    Wen, Yimin
    Fan, Zhigang
    2018 24TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2018, : 1349 - 1354
  • [25] Automatic online multi-source domain adaptation
    Renchunzi, Xie
    Pratama, Mahardhika
    INFORMATION SCIENCES, 2022, 582 : 480 - 494
  • [26] Multi-source domain adaptation for image classification
    Karimpour, Morvarid
    Noori Saray, Shiva
    Tahmoresnezhad, Jafar
    Pourmahmood Aghababa, Mohammad
    MACHINE VISION AND APPLICATIONS, 2020, 31 (06)
  • [27] Moment Matching for Multi-Source Domain Adaptation
    Peng, Xingchao
    Bai, Qinxun
    Xia, Xide
    Huang, Zijun
    Saenko, Kate
    Wang, Bo
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 1406 - 1415
  • [28] Multi-source domain adaptation for image classification
    Morvarid Karimpour
    Shiva Noori Saray
    Jafar Tahmoresnezhad
    Mohammad Pourmahmood Aghababa
    Machine Vision and Applications, 2020, 31
  • [29] Subspace Identification for Multi-Source Domain Adaptation
    Li, Zijian
    Cai, Ruichu
    Chen, Guangyi
    Sun, Boyang
    Hao, Zhifeng
    Zhang, Kun
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [30] Evidential combination of augmented multi-source of information based on domain adaptation
    Linqing Huang
    Zhunga Liu
    Quan Pan
    Jean Dezert
    Science China Information Sciences, 2020, 63