Pulling Up by the Causal Bootstraps: Causal Data Augmentation for Pre-training Debiasing

被引:1
|
作者
Gowda, Sindhu C. M. [1 ]
Joshi, Shalmali [2 ]
Zhang, Haoran [1 ]
Ghassemi, Marzyeh [3 ]
机构
[1] Univ Toronto, Vector Inst, Toronto, ON, Canada
[2] Harvard Univ, Cambridge, MA USA
[3] MIT, Cambridge, MA USA
基金
加拿大自然科学与工程研究理事会;
关键词
confounding bias; pre-training; debiasing; causal graphs; re-sampling; ALGORITHM; HEALTH; BIAS;
D O I
10.1145/3459637.3482380
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Machine learning models achieve state-of-the-art performance on many supervised learning tasks. However, prior evidence suggests that these models may learn to rely on "shortcut" biases or spurious correlations (intuitively, correlations that do not hold in the test as they hold in train) for good predictive performance. Such models cannot be trusted in deployment environments to provide accurate predictions. While viewing the problem from a causal lens is known to be useful, the seamless integration of causation techniques into machine learning pipelines remains cumbersome and expensive. In this work, we study and extend a causal pre-training debiasing technique called causal bootstrapping (CB) under five practical confounded-data generation-acquisition scenarios (with known and unknown confounding). Under these settings, we systematically investigate the effect of confounding bias on deep learning model performance, demonstrating their propensity to rely on shortcut biases when these biases are not properly accounted for. We demonstrate that such a causal pre-training technique can significantly outperform existing base practices to mitigate confounding bias on real-world domain generalization benchmarking tasks. This systematic investigation underlines the importance of accounting for the underlying data-generating mechanisms and fortifying data-preprocessing pipelines with a causal framework to develop methods robust to confounding biases.
引用
收藏
页码:606 / 616
页数:11
相关论文
共 50 条
  • [41] Incorporating Causal Graphical Prior Knowledge into Predictive Modeling via Simple Data Augmentation
    Teshima, Takeshi
    Sugiyama, Masashi
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, VOL 161, 2021, 161 : 86 - 96
  • [42] Towards Context-Aware Emotion Recognition Debiasing From a Causal Demystification Perspective via De-Confounded Training
    Yang, Dingkang
    Yang, Kun
    Kuang, Haopeng
    Chen, Zhaoyu
    Wang, Yuzheng
    Zhang, Lihua
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (12) : 10663 - 10680
  • [43] INGENIOUS: Using Informative Data Subsets for Efficient Pre-Training of Language Models
    Renduchintala, H. S. V. N. S. Kowndinya
    Killamsetty, Krishnateja
    Bhatia, Sumit
    Aggarwal, Milan
    Ramakrishnan, Ganesh
    Iyer, Rishabh
    Krishnamurthy, Balaji
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 6690 - 6705
  • [44] Unsupervised Pre-Training of Imbalanced Data for Identification of Wafer Map Defect Patterns
    Shon, Ho Sun
    Batbaatar, Erdenebileg
    Cho, Wan-Sup
    Choi, Seong Gon
    IEEE ACCESS, 2021, 9 : 52352 - 52363
  • [45] RPT: Toward Transferable Model on Heterogeneous Researcher Data via Pre-Training
    Qiao, Ziyue
    Fu, Yanjie
    Wang, Pengyang
    Xiao, Meng
    Ning, Zhiyuan
    Zhang, Denghui
    Du, Yi
    Zhou, Yuanchun
    IEEE TRANSACTIONS ON BIG DATA, 2023, 9 (01) : 186 - 199
  • [46] Unsupervised Pre-training Classifier Based on Restricted Boltzmann Machine with Imbalanced Data
    Fu, Xiaoyang
    SMART COMPUTING AND COMMUNICATION, SMARTCOM 2016, 2017, 10135 : 102 - 110
  • [47] Specialized Pre-Training of Neural Networks on Synthetic Data for Improving Paraphrase Generation
    O. H. Skurzhanskyi
    O. O. Marchenko
    A. V. Anisimov
    Cybernetics and Systems Analysis, 2024, 60 : 167 - 174
  • [48] Specialized Pre-Training of Neural Networks on Synthetic Data for Improving Paraphrase Generation
    Skurzhanskyi, O. H.
    Marchenko, O. O.
    Anisimov, A. V.
    CYBERNETICS AND SYSTEMS ANALYSIS, 2024, 60 (02) : 167 - 174
  • [49] Neural Grammatical Error Correction Systems with Unsupervised Pre-training on Synthetic Data
    Grundkiewicz, Roman
    Junczys-Dowmunt, Marcin
    Heafield, Kenneth
    INNOVATIVE USE OF NLP FOR BUILDING EDUCATIONAL APPLICATIONS, 2019, : 252 - 263
  • [50] On the Connection between Pre-training Data Diversity and Fine-tuning Robustness
    Ramanujan, Vivek
    Nguyen, Thao
    Oh, Sewoong
    Schmidt, Ludwig
    Farhadi, Ali
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,