Pulling Up by the Causal Bootstraps: Causal Data Augmentation for Pre-training Debiasing

被引:1
|
作者
Gowda, Sindhu C. M. [1 ]
Joshi, Shalmali [2 ]
Zhang, Haoran [1 ]
Ghassemi, Marzyeh [3 ]
机构
[1] Univ Toronto, Vector Inst, Toronto, ON, Canada
[2] Harvard Univ, Cambridge, MA USA
[3] MIT, Cambridge, MA USA
基金
加拿大自然科学与工程研究理事会;
关键词
confounding bias; pre-training; debiasing; causal graphs; re-sampling; ALGORITHM; HEALTH; BIAS;
D O I
10.1145/3459637.3482380
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Machine learning models achieve state-of-the-art performance on many supervised learning tasks. However, prior evidence suggests that these models may learn to rely on "shortcut" biases or spurious correlations (intuitively, correlations that do not hold in the test as they hold in train) for good predictive performance. Such models cannot be trusted in deployment environments to provide accurate predictions. While viewing the problem from a causal lens is known to be useful, the seamless integration of causation techniques into machine learning pipelines remains cumbersome and expensive. In this work, we study and extend a causal pre-training debiasing technique called causal bootstrapping (CB) under five practical confounded-data generation-acquisition scenarios (with known and unknown confounding). Under these settings, we systematically investigate the effect of confounding bias on deep learning model performance, demonstrating their propensity to rely on shortcut biases when these biases are not properly accounted for. We demonstrate that such a causal pre-training technique can significantly outperform existing base practices to mitigate confounding bias on real-world domain generalization benchmarking tasks. This systematic investigation underlines the importance of accounting for the underlying data-generating mechanisms and fortifying data-preprocessing pipelines with a causal framework to develop methods robust to confounding biases.
引用
收藏
页码:606 / 616
页数:11
相关论文
共 50 条
  • [1] Causal Document-Grounded Dialogue Pre-training
    Zhao, Yingxiu
    Yu, Bowen
    Yu, Haiyang
    Li, Bowen
    Li, Jinyang
    Wang, Chao
    Huang, Fei
    Li, Yongbin
    Zhang, Nevin L.
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 7160 - 7174
  • [2] Improving Commonsense Causal Reasoning by Adversarial Training and Data Augmentation
    Staliunaite, Ieva
    Gorinski, Philip John
    Iacobacci, Ignacio
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 13834 - 13842
  • [3] PACT: Perception-Action Causal Transformer for Autoregressive Robotics Pre-Training
    Bonatti, Rogerio
    Vemprala, Sai
    Ma, Shuang
    Frujeri, Felipe
    Chen, Shuhang
    Kapoor, Ashish
    2023 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, IROS, 2023, : 3621 - 3627
  • [4] Pre-trained Recommender Systems: A Causal Debiasing Perspective
    Lin, Ziqian
    Ding, Hao
    Hoang, Nghia Trong
    Kveton, Branislav
    Deoras, Anoop
    Wang, Hao
    PROCEEDINGS OF THE 17TH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING, WSDM 2024, 2024, : 424 - 433
  • [5] Early Rumor Detection based on Data Augmentation and Pre-training Transformer
    Hu, Yanjun
    Ju, Xinyi
    Ye, Zhousheng
    Khan, Sulaiman
    Yuan, Chengwu
    Lai, Qiran
    Liu, Junqiang
    2022 IEEE 12TH ANNUAL COMPUTING AND COMMUNICATION WORKSHOP AND CONFERENCE (CCWC), 2022, : 152 - 158
  • [6] Disentangled causal representation learning for debiasing recommendation with uniform data
    Yang, Xinxin
    Li, Xinwei
    Liu, Zhen
    Wang, Yannan
    Lu, Sibo
    Liu, Feng
    APPLIED INTELLIGENCE, 2024, 54 (08) : 6760 - 6775
  • [7] Learning Visual Robotic Control Efficiently with Contrastive Pre-training and Data Augmentation
    Zhan, Albert
    Zhao, Ruihan
    Pinto, Lerrel
    Abbeel, Pieter
    Laskin, Michael
    2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 4040 - 4047
  • [8] LuxemBERT: Simple and Practical Data Augmentation in Language Model Pre-Training for Luxembourgish
    Lothritz, Cedric
    Lebichot, Bertrand
    Allix, Kevin
    Veiber, Lisa
    Bissyande, Tegawende F.
    Klein, Jacques
    Boytsov, Andrey
    Goujon, Anne
    Lefebvre, Clement
    LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 5080 - 5089
  • [9] Improving Sequential Recommendations via Bidirectional Temporal Data Augmentation With Pre-Training
    Jiang, Juyong
    Zhang, Peiyan
    Luo, Yingtao
    Li, Chaozhuo
    Kim, Jae Boum
    Zhang, Kai
    Wang, Senzhang
    Kim, Sunghun
    Yu, Philip S.
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2025, 37 (05) : 2652 - 2664
  • [10] Event Camera Data Pre-training
    Yang, Yan
    Pan, Liyuan
    Liu, Liu
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 10665 - 10675