TEST-TIME DETECTION OF BACKDOOR TRIGGERS FOR POISONED DEEP NEURAL NETWORKS

被引:4
|
作者
Li, Xi [1 ]
Xiang, Zhen [1 ]
Miller, David J. [1 ]
Kesidis, George [1 ]
机构
[1] Penn State Univ, Sch EECS, Philadelphia, PA 19104 USA
关键词
adversarial learning; backdoor attack; Trojan attack; in-flight detection; image classification;
D O I
10.1109/ICASSP43922.2022.9746573
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Backdoor (Trojan) attacks are emerging threats against deep neural networks (DNN). A DNN being attacked will predict to an attacker-desired target class whenever a test sample from any source class is embedded with a backdoor pattern, while correctly classifying clean (attack-free) test samples. Existing backdoor defenses have shown success in detecting whether a DNN is attacked and in reverse-engineering the backdoor pattern in a "post-training" scenario: the defender has access to the DNN to be inspected and a small, clean dataset collected independently, but has no access to the (possibly poisoned) training set of the DNN. However, these defenses neither catch culprits in the act of triggering the backdoor mapping, nor mitigate the backdoor attack at testtime. In this paper, we propose an "in-flight" unsupervised defense against backdoor attacks on image classification that 1) detects use of a backdoor trigger at test-time; and 2) infers the class of origin (source class) for a detected trigger example. The effectiveness of our defense is demonstrated experimentally for a wide variety of DNN architectures, datasets, and backdoor attack configurations.
引用
收藏
页码:3333 / 3337
页数:5
相关论文
共 50 条
  • [41] TTANAD: Test-Time Augmentation for Network Anomaly Detection
    Cohen, Seffi
    Goldshlager, Niv
    Shapira, Bracha
    Rokach, Lior
    ENTROPY, 2023, 25 (05)
  • [42] Aleatoric uncertainty estimation with test-time augmentation for medical image segmentation with convolutional neural networks
    Wang, Guotai
    Li, Wenqi
    Aertsen, Michael
    Deprest, Jan
    Ourselin, Sebastien
    Vercauteren, Tom
    NEUROCOMPUTING, 2019, 338 : 34 - 45
  • [43] Test-Time Linear Out-of-Distribution Detection
    Fan, Ke
    Liu, Tong
    Qiu, Xingyu
    Wang, Yikai
    Huai, Lian
    Shangguan, Zeyu
    Gou, Shuang
    Liu, Fengjian
    Fu, Yuqian
    Fu, Yanwei
    Jiang, Xingqun
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 23752 - 23761
  • [44] Reverse Backdoor Distillation: Towards Online Backdoor Attack Detection for Deep Neural Network Models
    Yao, Zeming
    Zhang, Hangtao
    Guo, Yicheng
    Tian, Xin
    Peng, Wei
    Zou, Yi
    Zhang, Leo Yu
    Chen, Chao
    IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, 2024, 21 (06) : 5098 - 5111
  • [45] Deep Matching Prior: Test-Time Optimization for Dense Correspondence
    Hong, Sunghwan
    Kim, Seungryong
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 9887 - 9897
  • [46] Invisible Backdoor Attacks on Deep Neural Networks Via Steganography and Regularization
    Li, Shaofeng
    Xue, Minhui
    Zhao, Benjamin
    Zhu, Haojin
    Zhang, Xinpeng
    IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, 2021, 18 (05) : 2088 - 2105
  • [47] Detecting Backdoor Attacks via Class Difference in Deep Neural Networks
    Kwon, Hyun
    IEEE ACCESS, 2020, 8 : 191049 - 191056
  • [48] Compression-resistant backdoor attack against deep neural networks
    Mingfu Xue
    Xin Wang
    Shichang Sun
    Yushu Zhang
    Jian Wang
    Weiqiang Liu
    Applied Intelligence, 2023, 53 : 20402 - 20417
  • [49] Backdoor Attacks against Deep Neural Networks by Personalized Audio Steganography
    Liu, Peng
    Zhang, Shuyi
    Yao, Chuanjian
    Ye, Wenzhe
    Li, Xianxian
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 68 - 74
  • [50] SGBA: A stealthy scapegoat backdoor attack against deep neural networks
    He, Ying
    Shen, Zhili
    Xia, Chang
    Hua, Jingyu
    Tong, Wei
    Zhong, Sheng
    COMPUTERS & SECURITY, 2024, 136