Backdoor Defense via Deconfounded Representation Learning

被引:5
|
作者
Zhang, Zaixi [1 ,2 ]
Liu, Qi [1 ,2 ]
Wang, Zhicai [4 ]
Lu, Zepu [4 ]
Hu, Qingyong [3 ]
机构
[1] Univ Sci & Technol China, Sch Comp Sci & Technol, Anhui Prov Key Lab Big Data Anal & Applicat, Hefei, Peoples R China
[2] State Key Lab Cognit Intelligence, Hefei, Anhui, Peoples R China
[3] Hong Kong Univ Sci & Technol, Hong Kong, Peoples R China
[4] Univ Sci & Technol China, Hefei, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1109/CVPR52729.2023.01177
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNs) are recently shown to be vulnerable to backdoor attacks, where attackers embed hidden backdoors in the DNN model by injecting a few poisoned examples into the training dataset. While extensive efforts have been made to detect and remove backdoors from backdoored DNNs, it is still not clear whether a backdoor-free clean model can be directly obtained from poisoned datasets. In this paper, we first construct a causal graph to model the generation process of poisoned data and find that the backdoor attack acts as the confounder, which brings spurious associations between the input images and target labels, making the model predictions less reliable. Inspired by the causal understanding, we propose the Causality-inspired Backdoor Defense (CBD), to learn deconfounded representations for reliable classification. Specifically, a backdoored model is intentionally trained to capture the confounding effects. The other clean model dedicates to capturing the desired causal effects by minimizing the mutual information with the confounding representations from the backdoored model and employing a sample-wise re-weighting scheme. Extensive experiments on multiple benchmark datasets against 6 state-of-the-art attacks verify that our proposed defense method is effective in reducing backdoor threats while maintaining high accuracy in predicting benign samples. Further analysis shows that CBD can also resist potential adaptive attacks. The code is available at https://github.com/zaixizhang/CBD.
引用
收藏
页码:12228 / 12238
页数:11
相关论文
共 50 条
  • [1] VFLIP: A Backdoor Defense for Vertical Federated Learning via Identification and Purification
    Cho, Yungi
    Han, Woorim
    Yu, Miseon
    Lee, Younghan
    Bae, Ho
    Paek, Yunheung
    COMPUTER SECURITY-ESORICS 2024, PT IV, 2024, 14985 : 291 - 312
  • [2] FLPurifier: Backdoor Defense in Federated Learning via Decoupled Contrastive Training
    Zhang, Jiale
    Zhu, Chengcheng
    Sun, Xiaobing
    Ge, Chunpeng
    Chen, Bing
    Susilo, Willy
    Yu, Shui
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 4752 - 4766
  • [3] Backdoor Attack and Defense on Deep Learning: A Survey
    Bai, Yang
    Xing, Gaojie
    Wu, Hongyan
    Rao, Zhihong
    Ma, Chuan
    Wang, Shiping
    Liu, Xiaolei
    Zhou, Yimin
    Tang, Jiajia
    Huang, Kaijun
    Kang, Jiale
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2025, 12 (01): : 404 - 434
  • [4] BayBFed: Bayesian Backdoor Defense for Federated Learning
    Kumari, Kavita
    Rieger, Phillip
    Fereidooni, Hossein
    Jadliwala, Murtuza
    Sadeghi, Ahmad-Reza
    2023 IEEE SYMPOSIUM ON SECURITY AND PRIVACY, SP, 2023, : 737 - 754
  • [5] Defense against backdoor attack in federated learning
    Lu, Shiwei
    Li, Ruihu
    Liu, Wenbin
    Chen, Xuan
    COMPUTERS & SECURITY, 2022, 121
  • [6] Backdoor attacks on unsupervised graph representation learning
    Feng, Bingdao
    Jin, Di
    Wang, Xiaobao
    Cheng, Fangyu
    Guo, Siqi
    NEURAL NETWORKS, 2024, 180
  • [7] Textual Backdoor Defense via Poisoned Sample Recognition
    Shao, Kun
    Zhang, Yu
    Yang, Junan
    Liu, Hui
    APPLIED SCIENCES-BASEL, 2021, 11 (21):
  • [8] Backdoor Defense via Adaptively Splitting Poisoned Dataset
    Gao, Kuofeng
    Bai, Yang
    Gu, Jindong
    Yang, Yong
    Xia, Shu-Tao
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 4005 - 4014
  • [9] Provable Defense against Backdoor Policies in Reinforcement Learning
    Bharti, Shubham Kumar
    Zhang, Xuezhou
    Singla, Adish
    Zhu, Xiaojin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [10] Federated Learning Backdoor Defense Based on Watermark Integrity
    Hou, Yinjian
    Zhao, Yancheng
    Yao, Kaiqi
    2024 10TH INTERNATIONAL CONFERENCE ON BIG DATA AND INFORMATION ANALYTICS, BIGDIA 2024, 2024, : 288 - 294