Backdoor Defense via Deconfounded Representation Learning

被引:5
|
作者
Zhang, Zaixi [1 ,2 ]
Liu, Qi [1 ,2 ]
Wang, Zhicai [4 ]
Lu, Zepu [4 ]
Hu, Qingyong [3 ]
机构
[1] Univ Sci & Technol China, Sch Comp Sci & Technol, Anhui Prov Key Lab Big Data Anal & Applicat, Hefei, Peoples R China
[2] State Key Lab Cognit Intelligence, Hefei, Anhui, Peoples R China
[3] Hong Kong Univ Sci & Technol, Hong Kong, Peoples R China
[4] Univ Sci & Technol China, Hefei, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1109/CVPR52729.2023.01177
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNs) are recently shown to be vulnerable to backdoor attacks, where attackers embed hidden backdoors in the DNN model by injecting a few poisoned examples into the training dataset. While extensive efforts have been made to detect and remove backdoors from backdoored DNNs, it is still not clear whether a backdoor-free clean model can be directly obtained from poisoned datasets. In this paper, we first construct a causal graph to model the generation process of poisoned data and find that the backdoor attack acts as the confounder, which brings spurious associations between the input images and target labels, making the model predictions less reliable. Inspired by the causal understanding, we propose the Causality-inspired Backdoor Defense (CBD), to learn deconfounded representations for reliable classification. Specifically, a backdoored model is intentionally trained to capture the confounding effects. The other clean model dedicates to capturing the desired causal effects by minimizing the mutual information with the confounding representations from the backdoored model and employing a sample-wise re-weighting scheme. Extensive experiments on multiple benchmark datasets against 6 state-of-the-art attacks verify that our proposed defense method is effective in reducing backdoor threats while maintaining high accuracy in predicting benign samples. Further analysis shows that CBD can also resist potential adaptive attacks. The code is available at https://github.com/zaixizhang/CBD.
引用
收藏
页码:12228 / 12238
页数:11
相关论文
共 50 条
  • [41] DLP: towards active defense against backdoor attacks with decoupled learning process
    Ying, Zonghao
    Wu, Bin
    CYBERSECURITY, 2023, 6 (01)
  • [42] Deconfounded recommendation via causal intervention
    Yu, Dianer
    Li, Qian
    Wang, Xiangmeng
    Xu, Guandong
    NEUROCOMPUTING, 2023, 529 : 128 - 139
  • [43] Survey of Textual Backdoor Attack and Defense
    Zheng M.
    Lin Z.
    Liu Z.
    Fu P.
    Wang W.
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2024, 61 (01): : 221 - 242
  • [44] Identifying Backdoor Attacks in Federated Learning via Anomaly Detection
    Mi, Yuxi
    Sun, Yiheng
    Guan, Jihong
    Zhou, Shuigeng
    WEB AND BIG DATA, PT III, APWEB-WAIM 2023, 2024, 14333 : 111 - 126
  • [45] Contrastive Neuron Pruning for Backdoor Defense
    Feng, Yu
    Ma, Benteng
    Liu, Dongnan
    Zhang, Yanning
    Cai, Weidong
    Xia, Yong
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2025, 34 : 1234 - 1245
  • [46] ACTSS: Input Detection Defense against Backdoor Attacks via Activation Subset Scanning
    Xuan, Yuexin
    Chen, Xiaojun
    Zhao, Zhendong
    Ding, Yangyang
    Lv, Jianming
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [47] BaFFLe: Backdoor Detection via Feedback -based Federated Learning
    Andreina, Sebastien
    Marson, Giorgia Azzurra
    Moellering, Helen
    Karame, Ghassan
    2021 IEEE 41ST INTERNATIONAL CONFERENCE ON DISTRIBUTED COMPUTING SYSTEMS (ICDCS 2021), 2021, : 852 - 863
  • [48] TRAINING SET CLEANSING OF BACKDOOR POISONING BY SELF-SUPERVISED REPRESENTATION LEARNING
    Wang, Hang
    Karami, Sahar
    Dia, Ousmane
    Ritter, Hippolyt
    Emamjomeh-Zadeh, Ehsan
    Chen, Jiahui
    Xiang, Zhen
    Miller, David J.
    Kesidis, George
    arXiv, 2022,
  • [49] Training Set Cleansing of Backdoor Poisoning by Self-Supervised Representation Learning
    Wang, Hang
    Karami, Sahar
    Dia, Ousmane
    Ritter, Hippolyt
    Emamjomeh-Zadeh, Ehsan
    Chen, Jiahui
    Xiang, Zhen
    Miller, David J.
    Kesidis, George
    ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings, 2023,
  • [50] Black-box Backdoor Defense via Zero-shot Image Purification
    Shi, Yucheng
    Du, Mengnan
    Wu, Xuansheng
    Guan, Zihan
    Sun, Jin
    Liu, Ninghao
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,