Out-of-Distribution Generalization With Causal Feature Separation

被引:3
|
作者
Wang, Haotian [1 ,2 ]
Kuang, Kun [3 ]
Lan, Long [1 ,2 ]
Wang, Zige [4 ]
Huang, Wanrong [1 ,2 ]
Wu, Fei [5 ]
Yang, Wenjing [1 ,2 ]
机构
[1] Natl Univ Def Technol, Inst Quantum Informat, Changsha 410073, Peoples R China
[2] Natl Univ Def Technol, Coll Comp Sci & Technol, State Key Lab High Performance Comp, Changsha 410073, Peoples R China
[3] Zhejiang Univ, Coll Comp Sci & Technol, Key Lab Corneal Dis Res Zhejiang Prov, Hangzhou 310027, Zhejiang, Peoples R China
[4] Peking Univ, Beijing 100871, Peoples R China
[5] Zhejiang Univ, Inst Artificial Intelligence, Shanghai Inst Adv Study, Shanghai AI Lab, Hangzhou 310027, Zhejiang, Peoples R China
基金
中国国家自然科学基金;
关键词
Causal features separation; OOD generalization; selection bias; stable prediction; PREDICTION; SELECTION;
D O I
10.1109/TKDE.2023.3312255
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Driven by empirical risk minimization, machine learning algorithm tends to exploit subtle statistical correlations existing in the training environment for prediction, while the spurious correlations are unstable across environments, leading to poor generalization performance. Accordingly, the problem of the Out-of-distribution (OOD) generalization aims to exploit an invariant/stable relationship between features and outcomes that generalizes well on all possible environments. To address the spurious correlation induced by the selection bias, in this article, we propose a novel Clique-based Causal Feature Separation (CCFS) algorithm by explicitly incorporating the causal structure to identify causal features of outcome for OOD generalization. Specifically, the proposed CCFS algorithm identifies the largest clique in the learned causal skeleton. Theoretically, we guarantee that either the largest clique or the rest of the causal skeleton is exactly the set of all causal features of the outcome. Finally, we separate the causal features from the non-causal ones with a sample-reweighting decorrelator for OOD prediction. Extensive experiments validate the effectiveness of the proposed CCFS method on both causal feature identification and OOD generalization tasks.
引用
收藏
页码:1758 / 1772
页数:15
相关论文
共 50 条
  • [21] Out-of-Distribution Generalization via Risk Extrapolation
    Krueger, David
    Caballero, Ethan
    Jacobsen, Joern-Henrik
    Zhang, Amy
    Binas, Jonathan
    Zhang, Dinghuai
    Le Priol, Remi
    Courville, Aaron
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [22] Out-of-distribution generalization for learning quantum dynamics
    Matthias C. Caro
    Hsin-Yuan Huang
    Nicholas Ezzell
    Joe Gibbs
    Andrew T. Sornborger
    Lukasz Cincio
    Patrick J. Coles
    Zoë Holmes
    Nature Communications, 14
  • [23] Towards a Theoretical Framework of Out-of-Distribution Generalization
    Ye, Haotian
    Xie, Chuanlong
    Cai, Tianle
    Li, Ruichen
    Li, Zhenguo
    Wang, Liwei
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [24] Multi-Instance Causal Representation Learning for Instance Label Prediction and Out-of-Distribution Generalization
    Zhang, Weijia
    Zhang, Xuanhui
    Deng, Han-Wen
    Zhang, Min-Ling
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [25] Causal Representation Learning for Out-of-Distribution Recommendation
    Wang, Wenjie
    Lin, Xinyu
    Feng, Fuli
    He, Xiangnan
    Lin, Min
    Chua, Tat-Seng
    PROCEEDINGS OF THE ACM WEB CONFERENCE 2022 (WWW'22), 2022, : 3562 - 3571
  • [26] DIVE: Subgraph Disagreement for Graph Out-of-Distribution Generalization
    Sun, Xin
    Wang, Liang
    Liu, Qiang
    Wu, Shu
    Wang, Zilei
    Wang, Liang
    PROCEEDINGS OF THE 30TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2024, 2024, : 2794 - 2805
  • [27] Toward Out-of-Distribution Generalization Through Inductive Biases
    Moruzzi, Caterina
    PHILOSOPHY AND THEORY OF ARTIFICIAL INTELLIGENCE 2021, 2022, 63 : 57 - 66
  • [28] Discovering causally invariant features for out-of-distribution generalization
    Wang, Yujie
    Yu, Kui
    Xiang, Guodu
    Cao, Fuyuan
    Liang, Jiye
    PATTERN RECOGNITION, 2024, 150
  • [29] Verifying the Generalization of Deep Learning to Out-of-Distribution Domains
    Amir, Guy
    Maayan, Osher
    Zelazny, Tom
    Katz, Guy
    Schapira, Michael
    JOURNAL OF AUTOMATED REASONING, 2024, 68 (03)
  • [30] Graph Out-of-Distribution Generalization With Controllable Data Augmentation
    Lu, Bin
    Zhao, Ze
    Gan, Xiaoying
    Liang, Shiyu
    Fu, Luoyi
    Wang, Xinbing
    Zhou, Chenghu
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (11) : 6317 - 6329