Mixture-of-Experts Learner for Single Long-Tailed Domain Generalization

被引:6
|
作者
Wang, Mengzhu [1 ]
Yuan, Jianlong [1 ]
Wang, Zhibin [1 ]
机构
[1] Alibaba Grp, Beijing, Peoples R China
关键词
Domain Generalization; Mixture-of-Experts Learner; Saliency Map; Mutual Learning;
D O I
10.1145/3581783.3611871
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Domain generalization (DG) refers to the task of training a model on multiple source domains and test it on a different target domain with different distribution. In this paper, we address a more challenging and realistic scenario known as Single Long-Tailed Domain Generalization, where only one source domain is available and the minority class in this domain has an abundance of instances in other domains. To tackle this task, we propose a novel approach called Mixture-of-Experts Learner for Single Long-Tailed Domain Generalization (MoEL), which comprises two key strategies. The first strategy is a simple yet effective data augmentation technique that leverages saliency maps to identify important regions on the original images and preserves these regions during augmentation. The second strategy is a new skill-diverse expert learning approach that trains multiple experts from a single long-tailed source domain and leverages mutual learning to aggregate their learned knowledge for the unknown target domain. We evaluate our method on various benchmark datasets, including Digits-DG, CIFAR-10-C, PACS, and DomainNet, and demonstrate its superior performance compared to previous single domain generalization methods. Additionally, the ablation study is also conducted to illustrate the inner workings of our approach.
引用
收藏
页码:290 / 299
页数:10
相关论文
共 43 条
  • [21] Improving Generalization and Personalization in Long-Tailed Federated Learning via Classifier Retraining
    Li, Yuhang
    Liu, Tong
    Shen, Wenfeng
    Cui, Yangguang
    Lu, Weijia
    EURO-PAR 2024: PARALLEL PROCESSING, PART II, EURO-PAR 2024, 2024, 14802 : 408 - 423
  • [22] Meta-DMoE: Adapting to Domain Shift by Meta-Distillation from Mixture-of-Experts
    Zhong, Tao
    Chi, Zhixiang
    Gu, Li
    Wang, Yang
    Yu, Yuanhao
    Tang, Jin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [23] DESIRE-ME: Domain-Enhanced Supervised Information Retrieval Using Mixture-of-Experts
    Kasela, Pranav
    Pasi, Gabriella
    Perego, Raffaele
    Tonellotto, Nicola
    ADVANCES IN INFORMATION RETRIEVAL, ECIR 2024, PT II, 2024, 14609 : 111 - 125
  • [24] MDCS: More Diverse Experts with Consistency Self-distillation for Long-tailed Recognition
    Zhao, Qihao
    Jiang, Chen
    Hu, Wei
    Zhang, Fan
    Liu, Jun
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 11563 - 11574
  • [25] Multi-Domain Long-Tailed Learning: Challenges, Progress, and Prospects
    Fu, Panpan
    Yusof, Umi Kalsom
    IEEE ACCESS, 2024, 12 : 129528 - 129540
  • [26] ACE: Ally Complementary Experts for Solving Long-Tailed Recognition in One-Shot
    Cai, Jiarui
    Wang, Yizhou
    Hwang, Jenq-Neng
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 112 - 121
  • [27] Global and Local Mixture Consistency Cumulative Learning for Long-tailed Visual Recognitions
    Du, Fei
    Yang, Peng
    Jia, Qi
    Nan, Fengtao
    Chen, Xiaoting
    Yang, Yun
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 15814 - 15823
  • [28] Enhancing long-tailed classification via multi-strategy weighted experts with hybrid distillation
    Zeng, Wu
    Xiao, Zhengying
    MULTIMEDIA SYSTEMS, 2025, 31 (01)
  • [29] A mixture-of-experts deep generative model for integrated analysis of single-cell multiomics data
    Minoura, Kodai
    Abe, Ko
    Nam, Hyunha
    Nishikawa, Hiroyoshi
    Shimamura, Teppei
    CELL REPORTS METHODS, 2021, 1 (05):
  • [30] Self-Supervised Aggregation of Diverse Experts for Test-Agnostic Long-Tailed Recognition
    Zhang, Yifan
    Hooi, Bryan
    Hong, Lanqing
    Feng, Jiashi
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,