Mixture-of-Experts Learner for Single Long-Tailed Domain Generalization

被引:6
|
作者
Wang, Mengzhu [1 ]
Yuan, Jianlong [1 ]
Wang, Zhibin [1 ]
机构
[1] Alibaba Grp, Beijing, Peoples R China
关键词
Domain Generalization; Mixture-of-Experts Learner; Saliency Map; Mutual Learning;
D O I
10.1145/3581783.3611871
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Domain generalization (DG) refers to the task of training a model on multiple source domains and test it on a different target domain with different distribution. In this paper, we address a more challenging and realistic scenario known as Single Long-Tailed Domain Generalization, where only one source domain is available and the minority class in this domain has an abundance of instances in other domains. To tackle this task, we propose a novel approach called Mixture-of-Experts Learner for Single Long-Tailed Domain Generalization (MoEL), which comprises two key strategies. The first strategy is a simple yet effective data augmentation technique that leverages saliency maps to identify important regions on the original images and preserves these regions during augmentation. The second strategy is a new skill-diverse expert learning approach that trains multiple experts from a single long-tailed source domain and leverages mutual learning to aggregate their learned knowledge for the unknown target domain. We evaluate our method on various benchmark datasets, including Digits-DG, CIFAR-10-C, PACS, and DomainNet, and demonstrate its superior performance compared to previous single domain generalization methods. Additionally, the ablation study is also conducted to illustrate the inner workings of our approach.
引用
收藏
页码:290 / 299
页数:10
相关论文
共 43 条
  • [31] Bilinear-experts network with self-adaptive sampler for long-tailed visual recognition
    Wang, Qin
    Kwong, Sam
    Wang, Xizhao
    NEUROCOMPUTING, 2025, 633
  • [32] Cross-Domain Empirical Risk Minimization for Unbiased Long-Tailed Classification
    Zhu, Beier
    Niu, Yulei
    Hua, Xian-Sheng
    Zhang, Hanwang
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 3589 - 3597
  • [33] M3oE: Multi-Domain Multi-Task Mixture-of-Experts Recommendation Framework
    Zhang, Zijian
    Liu, Shuchang
    Yu, Jiaao
    Cai, Qingpeng
    Zhao, Xiangyu
    Zhang, Chunxu
    Liu, Ziru
    Liu, Qidong
    Zhao, Hongwei
    Hu, Lantao
    Jiang, Peng
    Gai, Kun
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 893 - 902
  • [34] Three Heads Are Better than One: Complementary Experts for Long-Tailed Semi-supervised Learning
    Ma, Chengcheng
    Elezi, Ismail
    Deng, Jiankang
    Dong, Weiming
    Xu, Changsheng
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 13, 2024, : 14229 - 14237
  • [35] Collapse of Langmuir monolayer formed by the mixture of short- and long-tailed fatty acid molecules
    Sah, Bijay K.
    Kundu, Sarathi
    COLLOID AND INTERFACE SCIENCE COMMUNICATIONS, 2020, 36
  • [36] Accelerating Mixture-of-Experts language model inference via plug-and-play lookahead gate on a single GPU
    Ou, Jie
    Chen, Yueming
    Xiong, Buyao
    Wang, Zhaokun
    Tian, Wenhong
    COMPUTER STANDARDS & INTERFACES, 2025, 94
  • [37] Mixture-of-Experts Variational Autoencoder for clustering and generating from similarity-based representations on single cell data
    Kopf, Andreas
    Fortuin, Vincent
    Somnath, Vignesh Ram
    Claassen, Manfred
    PLOS COMPUTATIONAL BIOLOGY, 2021, 17 (06)
  • [38] Leveraging Semisupervised Learning for Domain Adaptation: Enhancing Safety at Construction Sites through Long-Tailed Object Detection
    Tran, Dai Quoc
    Jeon, Yuntae
    Aboah, Armstrong
    Bak, Jinyeong
    Park, Minsoo
    Park, Seunghee
    JOURNAL OF CONSTRUCTION ENGINEERING AND MANAGEMENT, 2025, 151 (01)
  • [39] Improving Mortality Prediction in ICU by Learning Long-Tailed Population of Patients Through Graph-Domain Aggregation
    Shin, Yunseob
    Tae, Yunwon
    Lee, Yeha
    IEEE ACCESS, 2022, 10 : 84405 - 84416
  • [40] SememeASR: Boosting Performance of End-to-End Speech Recognition against Domain and Long-Tailed Data Shift with Sememe Semantic Knowledge
    Zhu, Jiaxu
    Song, Changhe
    Wu, Zhiyong
    Meng, Helen
    INTERSPEECH 2023, 2023, : 3272 - 3276