DecAug: Out-of-Distribution Generalization via Decomposed Feature Representation and Semantic Augmentation

被引:0
|
作者
Bai, Haoyue [1 ,2 ]
Sun, Rui [2 ]
Hong, Lanqing [2 ]
Zhou, Fengwei [2 ]
Ye, Nanyang [3 ]
Ye, Han-Jia [4 ]
Chan, S-H Gary [1 ]
Li, Zhenguo [2 ]
机构
[1] Hong Kong Univ Sci & Technol, Hong Kong, Peoples R China
[2] Huawei Noahs Ark Lab, Hong Kong, Peoples R China
[3] Shanghai Jiao Tong Univ, Shanghai, Peoples R China
[4] Nanjing Univ, Nanjing, Jiangsu, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
While deep learning demonstrates its strong ability to handle independent and identically distributed (IID) data, it often suffers from out-of-distribution (OoD) generalization, where the test data come from another distribution (w.r.t. the training one). Designing a general OoD generalization framework for a wide range of applications is challenging, mainly due to different kinds of distribution shifts in the real world, such as the shift across domains or the extrapolation of correlation. Most of the previous approaches can only solve one specific distribution shift, leading to unsatisfactory performance when applied to various OoD benchmarks. In this work, we propose DecAug, a novel decomposed feature representation and semantic augmentation approach for OoD generalization. Specifically, DecAug disentangles the category-related and context-related features by orthogonalizing the two gradients (w.r.t. intermediate features) of losses for predicting category and context labels, where category-related features contain causal information of the target object, while context-related features cause distribution shifts between training and test data. Furthermore, we perform gradient-based augmentation on context-related features to improve the robustness of learned representations. Experimental results show that DecAug outperforms other state-of-the-art methods on various OoD datasets, which is among the very few methods that can deal with different types of OoD generalization challenges.
引用
收藏
页码:6705 / 6713
页数:9
相关论文
共 50 条
  • [31] Toward Out-of-Distribution Generalization Through Inductive Biases
    Moruzzi, Caterina
    PHILOSOPHY AND THEORY OF ARTIFICIAL INTELLIGENCE 2021, 2022, 63 : 57 - 66
  • [32] DIVE: Subgraph Disagreement for Graph Out-of-Distribution Generalization
    Sun, Xin
    Wang, Liang
    Liu, Qiang
    Wu, Shu
    Wang, Zilei
    Wang, Liang
    PROCEEDINGS OF THE 30TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2024, 2024, : 2794 - 2805
  • [33] Verifying the Generalization of Deep Learning to Out-of-Distribution Domains
    Amir, Guy
    Maayan, Osher
    Zelazny, Tom
    Katz, Guy
    Schapira, Michael
    JOURNAL OF AUTOMATED REASONING, 2024, 68 (03)
  • [34] Discovering causally invariant features for out-of-distribution generalization
    Wang, Yujie
    Yu, Kui
    Xiang, Guodu
    Cao, Fuyuan
    Liang, Jiye
    PATTERN RECOGNITION, 2024, 150
  • [35] Multi-Instance Causal Representation Learning for Instance Label Prediction and Out-of-Distribution Generalization
    Zhang, Weijia
    Zhang, Xuanhui
    Deng, Han-Wen
    Zhang, Min-Ling
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [36] Probing out-of-distribution generalization in machine learning for materials
    Li, Kangming
    Rubungo, Andre Niyongabo
    Lei, Xiangyun
    Persaud, Daniel
    Choudhary, Kamal
    Decost, Brian
    Dieng, Adji Bousso
    Hattrick-Simpers, Jason
    COMMUNICATIONS MATERIALS, 2025, 6 (01)
  • [37] Tackling Domain Generalization for Out-of-Distribution Endoscopic Imaging
    Ali Teevno, Mansoor
    Ochoa-Ruiz, Gilberto
    Ali, Sharib
    MACHINE LEARNING IN MEDICAL IMAGING, PT II, MLMI 2024, 2025, 15242 : 43 - 52
  • [38] RetroOOD: Understanding Out-of-Distribution Generalization in Retrosynthesis Prediction
    Yu, Yemin
    Yuan, Luotian
    Wei, Ying
    Gao, Hanyu
    Wu, Fei
    Wang, Zhihua
    Ye, Xinhai
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 1, 2024, : 374 - 382
  • [39] Can Subnetwork Structure be the Key to Out-of-Distribution Generalization?
    Zhang, Dinghuai
    Ahuja, Kartik
    Xu, Yilun
    Wang, Yisen
    Courville, Aaron
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [40] Face Reconstruction Transfer Attack as Out-of-Distribution Generalization
    June, Yoon Gyo
    Park, Jaewoo
    Dong, Xingbo
    Park, Hojin
    Teoh, Andrew Beng Jin
    Camps, Octavia
    COMPUTER VISION - ECCV 2024, PT LXXV, 2025, 15133 : 396 - 413