Can We Mitigate Backdoor Attack Using Adversarial Detection Methods?

被引:3
|
作者
Jin, Kaidi [1 ]
Zhang, Tianwei [2 ]
Shen, Chao [1 ]
Chen, Yufei [1 ,3 ]
Fan, Ming [1 ,4 ]
Lin, Chenhao [1 ]
Liu, Ting [1 ]
机构
[1] Xi An Jiao Tong Univ, Fac Elect & Infomat Engn, Xian 710049, Peoples R China
[2] Nanyang Technol Univ, Sch Comp Sci, Engn, Singapore 639798, Singapore
[3] City Univ Hong Kong, Hong Kong, Peoples R China
[4] Chongqing Univ, Key Lab Dependable Serv Comp Cyber Phys Soc, Minist Educ China, Chongqing 400044, Peoples R China
基金
中国国家自然科学基金;
关键词
Adversarial attacks; backdoor attacks; deep neural networks; robustness;
D O I
10.1109/TDSC.2022.3194642
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Deep Neural Networks are well known to be vulnerable to adversarial attacks and backdoor attacks, where minor modifications on the input are able to mislead the models to give wrong results. Although defenses against adversarial attacks have been widely studied, investigation on mitigating backdoor attacks is still at an early stage. It is unknown whether there are any connections and common characteristics between the defenses against these two attacks. We conduct comprehensive studies on the connections between adversarial examples and backdoor examples of Deep Neural Networks to seek to answer the question: can we detect backdoor using adversarial detection methods. Our insights are based on the observation that both adversarial examples and backdoor examples have anomalies during the inference process, highly distinguishable from benign samples. As a result, we revise four existing adversarial defense methods for detecting backdoor examples. Extensive evaluations indicate that these approaches provide reliable protection against backdoor attacks, with a higher accuracy than detecting adversarial examples. These solutions also reveal the relations of adversarial examples, backdoor examples and normal samples in model sensitivity, activation space and feature space. This is able to enhance our understanding about the inherent features of these two attacks and the defense opportunities.
引用
收藏
页码:2867 / 2881
页数:15
相关论文
共 50 条
  • [1] STEALTHY BACKDOOR ATTACK WITH ADVERSARIAL TRAINING
    Feng, Le
    Li, Sheng
    Qian, Zhenxing
    Zhang, Xinpeng
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 2969 - 2973
  • [2] Attack as Detection: Using Adversarial Attack Methods to Detect Abnormal Examples
    Zhao, Zhe
    Chen, Guangke
    Liu, Tong
    Li, Taishan
    Song, Fu
    Wang, Jingyi
    Sun, Jun
    ACM TRANSACTIONS ON SOFTWARE ENGINEERING AND METHODOLOGY, 2024, 33 (03)
  • [3] AdvDoor: Adversarial Backdoor Attack of Deep Learning System
    Zhang, Quan
    Ding, Yifeng
    Tian, Yongqiang
    Guo, Jianmin
    Yuan, Min
    Jiang, Yu
    ISSTA '21: PROCEEDINGS OF THE 30TH ACM SIGSOFT INTERNATIONAL SYMPOSIUM ON SOFTWARE TESTING AND ANALYSIS, 2021, : 127 - 138
  • [4] Evil vs evil: using adversarial examples to against backdoor attack in federated learning
    Liu, Tao
    Li, Mingjun
    Zheng, Haibin
    Ming, Zhaoyan
    Chen, Jinyin
    MULTIMEDIA SYSTEMS, 2023, 29 (02) : 553 - 568
  • [5] Evil vs evil: using adversarial examples to against backdoor attack in federated learning
    Tao Liu
    Mingjun Li
    Haibin Zheng
    Zhaoyan Ming
    Jinyin Chen
    Multimedia Systems, 2023, 29 : 553 - 568
  • [6] APBAM: Adversarial perturbation-driven backdoor attack in multimodal learning
    Zhang, Shaobo
    Chen, Wenli
    Li, Xiong
    Liu, Qin
    Wang, Guojun
    INFORMATION SCIENCES, 2025, 700
  • [7] Camouflage Backdoor Attack against Pedestrian Detection
    Wu, Yalun
    Gu, Yanfeng
    Chen, Yuanwan
    Cui, Xiaoshu
    Li, Qiong
    Xiang, Yingxiao
    Tong, Endong
    Li, Jianhua
    Han, Zhen
    Liu, Jiqiang
    APPLIED SCIENCES-BASEL, 2023, 13 (23):
  • [8] Federated Learning Backdoor Attack Scheme Based on Generative Adversarial Network
    Chen D.
    Fu A.
    Zhou C.
    Chen Z.
    Fu, Anmin (fuam@njust.edu.cn); Fu, Anmin (fuam@njust.edu.cn), 1600, Science Press (58): : 2364 - 2373
  • [9] A Synergetic Attack against Neural Network Classifiers combining Backdoor and Adversarial Examples
    Liu, Guanxiong
    Khalil, Issa
    Khreishah, Abdallah
    Phan, NhatHai
    2021 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2021, : 834 - 846
  • [10] Adversarial Backdoor Attack by Naturalistic Data Poisoning on Trajectory Prediction in Autonomous Driving
    Pourkeshavarz, Mozhgan
    Sabokrou, Mohammad
    Rasouli, Amir
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 14885 - 14894