Backdoor Defense via Deconfounded Representation Learning

被引:5
|
作者
Zhang, Zaixi [1 ,2 ]
Liu, Qi [1 ,2 ]
Wang, Zhicai [4 ]
Lu, Zepu [4 ]
Hu, Qingyong [3 ]
机构
[1] Univ Sci & Technol China, Sch Comp Sci & Technol, Anhui Prov Key Lab Big Data Anal & Applicat, Hefei, Peoples R China
[2] State Key Lab Cognit Intelligence, Hefei, Anhui, Peoples R China
[3] Hong Kong Univ Sci & Technol, Hong Kong, Peoples R China
[4] Univ Sci & Technol China, Hefei, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1109/CVPR52729.2023.01177
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNs) are recently shown to be vulnerable to backdoor attacks, where attackers embed hidden backdoors in the DNN model by injecting a few poisoned examples into the training dataset. While extensive efforts have been made to detect and remove backdoors from backdoored DNNs, it is still not clear whether a backdoor-free clean model can be directly obtained from poisoned datasets. In this paper, we first construct a causal graph to model the generation process of poisoned data and find that the backdoor attack acts as the confounder, which brings spurious associations between the input images and target labels, making the model predictions less reliable. Inspired by the causal understanding, we propose the Causality-inspired Backdoor Defense (CBD), to learn deconfounded representations for reliable classification. Specifically, a backdoored model is intentionally trained to capture the confounding effects. The other clean model dedicates to capturing the desired causal effects by minimizing the mutual information with the confounding representations from the backdoored model and employing a sample-wise re-weighting scheme. Extensive experiments on multiple benchmark datasets against 6 state-of-the-art attacks verify that our proposed defense method is effective in reducing backdoor threats while maintaining high accuracy in predicting benign samples. Further analysis shows that CBD can also resist potential adaptive attacks. The code is available at https://github.com/zaixizhang/CBD.
引用
收藏
页码:12228 / 12238
页数:11
相关论文
共 50 条
  • [21] Backdoor Attack and Defense in Asynchronous Federated Learning for Multiple Unmanned Vehicles
    Wang, Kehao
    Zhang, Hao
    2024 3RD CONFERENCE ON FULLY ACTUATED SYSTEM THEORY AND APPLICATIONS, FASTA 2024, 2024, : 843 - 847
  • [22] Backdoor Attack Defense Method for Federated Learning Based on Model Watermarking
    Guo J.-J.
    Liu J.-Z.
    Ma Y.
    Liu Z.-Q.
    Xiong Y.-P.
    Miao K.
    Li J.-X.
    Ma J.-F.
    Jisuanji Xuebao/Chinese Journal of Computers, 2024, 47 (03): : 662 - 676
  • [23] Knowledge Distillation Based Defense for Audio Trigger Backdoor in Federated Learning
    Chen, Yu-Wen
    Ke, Bo-Hsu
    Chen, Bo-Zhong
    Chiu, Si-Rong
    Tu, Chun-Wei
    Kuo, Jian-Jhih
    IEEE CONFERENCE ON GLOBAL COMMUNICATIONS, GLOBECOM, 2023, : 4271 - 4276
  • [24] Successive Interference Cancellation Based Defense for Trigger Backdoor in Federated Learning
    Chen, Yu-Wen
    Ke, Bo-Hsu
    Chen, Bo-Zhong
    Chiu, Si-Rong
    Tu, Chun-Wei
    Kuo, Jian-Jhih
    ICC 2023-IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS, 2023, : 26 - 32
  • [25] DAGUARD: distributed backdoor attack defense scheme under federated learning
    Yu S.
    Chen Z.
    Chen Z.
    Liu X.
    Tongxin Xuebao/Journal on Communications, 2023, 44 (05): : 110 - 122
  • [26] Towards robustness evaluation of backdoor defense on quantized deep learning models
    Zhu, Yifan
    Peng, Huaibing
    Fu, Anmin
    Yang, Wei
    Ma, Hua
    Al-Sarawi, Said F.
    Abbott, Derek
    Gao, Yansong
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 255
  • [27] A Stability-Enhanced Dynamic Backdoor Defense in Federated Learning for IIoT
    Ma, Zhixuan
    Gao, Haichang
    Li, Shangwen
    Wang, Ping
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2024, 20 (11) : 12513 - 12522
  • [28] FederatedReverse: A Detection and Defense Method Against Backdoor Attacks in Federated Learning
    Zhao, Chen
    Wen, Yu
    Li, Shuailou
    Liu, Fucheng
    Meng, Dan
    PROCEEDINGS OF THE 2021 ACM WORKSHOP ON INFORMATION HIDING AND MULTIMEDIA SECURITY, IH&MMSEC 2021, 2021, : 51 - 62
  • [29] Backdoor Defense with Machine Unlearning
    Liu, Yang
    Fan, Mingyuan
    Chen, Cen
    Liu, Ximeng
    Ma, Zhuo
    Wang, Li
    Ma, Jianfeng
    IEEE CONFERENCE ON COMPUTER COMMUNICATIONS (IEEE INFOCOM 2022), 2022, : 280 - 289
  • [30] OCIE: Augmenting model interpretability via Deconfounded Explanation-Guided Learning
    Dong, Liang
    Chen, Leiyang
    Zheng, Chengliang
    Fu, Zhongwang
    Zukaib, Umer
    Cui, Xiaohui
    Shen, Zhidong
    KNOWLEDGE-BASED SYSTEMS, 2024, 302