Transformer Based Multi-Source Domain Adaptation

被引:0
|
作者
Wright, Dustin [1 ]
Augenstein, Isabelle [1 ]
机构
[1] Univ Copenhagen, Dept Comp Sci, Copenhagen, Denmark
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In practical machine learning settings, the data on which a model must make predictions often come from a different distribution than the data it was trained on. Here, we investigate the problem of unsupervised multi-source domain adaptation, where a model is trained on labelled data from multiple source domains and must make predictions on a domain for which no labelled data has been seen. Prior work with CNNs and RNNs has demonstrated the benefit of mixture of experts, where the predictions of multiple domain expert classifiers are combined; as well as domain adversarial training, to induce a domain agnostic representation space. Inspired by this, we investigate how such methods can be effectively applied to large pretrained transformer models. We find that domain adversarial training has an effect on the learned representations of these models while having little effect on their performance, suggesting that large transformer-based models are already relatively robust across domains. Additionally, we show that mixture of experts leads to significant performance improvements by comparing several variants of mixing functions, including one novel mixture based on attention. Finally, we demonstrate that the predictions of large pretrained transformer based domain experts are highly homogenous, making it challenging to learn effective functions for mixing their predictions.
引用
收藏
页码:7963 / 7974
页数:12
相关论文
共 50 条
  • [1] Transformer-Based Multi-Source Domain Adaptation Without Source Data
    Li, Gang
    Wu, Chao
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [2] Contrastive transformer based domain adaptation for multi-source cross-domain sentiment classification
    Fu, Yanping
    Liu, Yun
    KNOWLEDGE-BASED SYSTEMS, 2022, 245
  • [3] Attention-Based Multi-Source Domain Adaptation
    Zuo, Yukun
    Yao, Hantao
    Xu, Changsheng
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 3793 - 3803
  • [4] Multi-source based approach for Visual Domain Adaptation
    Tiwari, Mrinalini
    Sanodiya, Rakesh Kumar
    Mathew, Jimson
    Saha, Sriparna
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [5] A survey of multi-source domain adaptation
    Sun, Shiliang
    Shi, Honglei
    Wu, Yuanbin
    INFORMATION FUSION, 2015, 24 : 84 - 92
  • [6] Multi-Source Distilling Domain Adaptation
    Zhao, Sicheng
    Wang, Guangzhi
    Zhang, Shanghang
    Gu, Yang
    Li, Yaxian
    Song, Zhichao
    Xu, Pengfei
    Hu, Runbo
    Chai, Hua
    Keutzer, Kurt
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 12975 - 12983
  • [7] BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION
    Sun, Shi-Liang
    Shi, Hong-Lei
    PROCEEDINGS OF 2013 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS (ICMLC), VOLS 1-4, 2013, : 24 - 28
  • [8] Multi-Source Survival Domain Adaptation
    Shaker, Ammar
    Lawrence, Carolin
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 8, 2023, : 9752 - 9762
  • [10] Wasserstein Barycenter for Multi-Source Domain Adaptation
    Montesuma, Eduardo Fernandes
    Mboula, Fred Maurice Ngole
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 16780 - 16788