MixStyle Neural Networks for Domain Generalization and Adaptation

被引:7
|
作者
Zhou, Kaiyang [1 ]
Yang, Yongxin [2 ]
Qiao, Yu [3 ,4 ]
Xiang, Tao [5 ]
机构
[1] Hong Kong Baptist Univ, Hong Kong, Peoples R China
[2] Queen Mary Univ London, London, England
[3] Chinese Acad Sci, Shanghai AI Lab, Shenzhen, Peoples R China
[4] Chinese Acad Sci, Shenzhen Inst Adv Technol, Shenzhen, Peoples R China
[5] Univ Surrey, Guildford, England
关键词
D O I
10.1007/s11263-023-01913-8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Neural networks do not generalize well to unseen data with domain shifts-a longstanding problem in machine learning and AI. To overcome the problem, we propose MixStyle, a simple plug-and-play, parameter-free module that can improve domain generalization performance without the need to collect more data or increase model capacity. The design of MixStyle is simple: it mixes the feature statistics of two random instances in a single forward pass during training. The idea is grounded by the finding from recent style transfer research that feature statistics capture image style information, which essentially defines visual domains. Therefore, mixing feature statistics can be seen as an efficient way to synthesize new domains in the feature space, thus achieving data augmentation. MixStyle is easy to implement with a few lines of code, does not require modification to training objectives, and can fit a variety of learning paradigms including supervised domain generalization, semi-supervised domain generalization, and unsupervised domain adaptation. Our experiments show that MixStyle can significantly boost out-of-distribution generalization performance across a wide range of tasks including image recognition, instance retrieval and reinforcement learning. The source code is released at https://github.com/KaiyangZhou/mixstyle-release.
引用
收藏
页码:822 / 836
页数:15
相关论文
共 50 条
  • [21] Generalization theory and generalization methods for neural networks
    Wei, Hai-Kun
    Xu, Si-Xin
    Song, Wen-Zhong
    Zidonghua Xuebao/Acta Automatica Sinica, 2001, 27 (06): : 806 - 815
  • [22] Domain Neural Adaptation
    Chen, Sentao
    Hong, Zijie
    Harandi, Mehrtash
    Yang, Xiaowei
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (11) : 8630 - 8641
  • [23] Domain generalization in rotating machinery fault diagnostics using deep neural networks
    Li, Xiang
    Zhang, Wei
    Ma, Hui
    Luo, Zhong
    Li, Xu
    NEUROCOMPUTING, 2020, 403 : 409 - 420
  • [24] Multimodal Emotion Recognition Method Based on Domain Generalization and Graph Neural Networks
    Xie, Jinbao
    Wang, Yulong
    Meng, Tianxin
    Tai, Jianqiao
    Zheng, Yueqian
    Varatnitski, Yury I.
    ELECTRONICS, 2025, 14 (05):
  • [25] Vision transformers in domain adaptation and domain generalization: a study of robustness
    Alijani, Shadi
    Fayyad, Jamil
    Najjaran, Homayoun
    Neural Computing and Applications, 2024, 36 (29) : 17979 - 18007
  • [26] On generalization in moment-based domain adaptation
    Zellinger, Werner
    Moser, Bernhard A.
    Saminger-Platz, Susanne
    ANNALS OF MATHEMATICS AND ARTIFICIAL INTELLIGENCE, 2021, 89 (3-4) : 333 - 369
  • [27] Style Normalization and Restitution for Domain Generalization and Adaptation
    Jin, Xin
    Lan, Cuiling
    Zeng, Wenjun
    Chen, Zhibo
    IEEE TRANSACTIONS ON MULTIMEDIA, 2021, 24 : 3636 - 3651
  • [28] Feature Diversification and Adaptation for Federated Domain Generalization
    Yang, Seunghan
    Choi, Seokeon
    Park, Hyunsin
    Choi, Sungha
    Chang, Simyung
    Yuri, Sungrack
    COMPUTER VISION - ECCV 2024, PT LXXII, 2025, 15130 : 52 - 70
  • [29] Open-world Domain Adaptation and Generalization
    Zhao, Sicheng
    Tao, Jianhua
    Ding, Guiguang
    PROCEEDINGS OF THE ACM TURING AWARD CELEBRATION CONFERENCE-CHINA 2024, ACM-TURC 2024, 2024, : 201 - 202
  • [30] Unified Deep Supervised Domain Adaptation and Generalization
    Motiian, Saeid
    Piccirilli, Marco
    Adjeroh, Donald A.
    Doretto, Gianfranco
    2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 5716 - 5726