TEST-TIME DETECTION OF BACKDOOR TRIGGERS FOR POISONED DEEP NEURAL NETWORKS

被引:4
|
作者
Li, Xi [1 ]
Xiang, Zhen [1 ]
Miller, David J. [1 ]
Kesidis, George [1 ]
机构
[1] Penn State Univ, Sch EECS, Philadelphia, PA 19104 USA
关键词
adversarial learning; backdoor attack; Trojan attack; in-flight detection; image classification;
D O I
10.1109/ICASSP43922.2022.9746573
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Backdoor (Trojan) attacks are emerging threats against deep neural networks (DNN). A DNN being attacked will predict to an attacker-desired target class whenever a test sample from any source class is embedded with a backdoor pattern, while correctly classifying clean (attack-free) test samples. Existing backdoor defenses have shown success in detecting whether a DNN is attacked and in reverse-engineering the backdoor pattern in a "post-training" scenario: the defender has access to the DNN to be inspected and a small, clean dataset collected independently, but has no access to the (possibly poisoned) training set of the DNN. However, these defenses neither catch culprits in the act of triggering the backdoor mapping, nor mitigate the backdoor attack at testtime. In this paper, we propose an "in-flight" unsupervised defense against backdoor attacks on image classification that 1) detects use of a backdoor trigger at test-time; and 2) infers the class of origin (source class) for a detected trigger example. The effectiveness of our defense is demonstrated experimentally for a wide variety of DNN architectures, datasets, and backdoor attack configurations.
引用
收藏
页码:3333 / 3337
页数:5
相关论文
共 50 条
  • [21] Hibernated Backdoor: A Mutual Information Empowered Backdoor Attack to Deep Neural Networks
    Ning, Rui
    Li, Jiang
    Xin, Chunsheng
    Wu, Hongyi
    Wang, Chonggang
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 10309 - 10318
  • [22] Patch Based Backdoor Attack on Deep Neural Networks
    Manna, Debasmita
    Tripathy, Somanath
    INFORMATION SYSTEMS SECURITY, ICISS 2024, 2025, 15416 : 422 - 440
  • [23] Detection of backdoor attacks using targeted universal adversarial perturbations for deep neural networks
    Qu, Yubin
    Huang, Song
    Chen, Xiang
    Wang, Xingya
    Yao, Yongming
    JOURNAL OF SYSTEMS AND SOFTWARE, 2024, 207
  • [24] Interpretability Derived Backdoor Attacks Detection in Deep Neural Networks: Work-in-Progress
    Wen, Xiangyu
    Jiang, Wei
    Zhan, Jinyu
    Wang, Xupeng
    He, Zhiyuan
    PROCEEDINGS OF THE 2020 INTERNATIONAL CONFERENCE ON EMBEDDED SOFTWARE (EMSOFT), 2020, : 13 - 14
  • [25] Backdoor Attack on Deep Neural Networks in Perception Domain
    Mo, Xiaoxing
    Zhang, Leo Yu
    Sun, Nan
    Luo, Wei
    Gao, Shang
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [26] On the Robustness of Backdoor-basedWatermarking in Deep Neural Networks
    Shafieinejad, Masoumeh
    Lukas, Nils
    Wang, Jiaqi
    Li, Xinda
    Kerschbaum, Florian
    PROCEEDINGS OF THE 2021 ACM WORKSHOP ON INFORMATION HIDING AND MULTIMEDIA SECURITY, IH&MMSEC 2021, 2021, : 177 - 188
  • [27] Adaptive Backdoor Attack against Deep Neural Networks
    He, Honglu
    Zhu, Zhiying
    Zhang, Xinpeng
    CMES-COMPUTER MODELING IN ENGINEERING & SCIENCES, 2023, 136 (03): : 2617 - 2633
  • [28] Automatic Brain Tumor Segmentation Using Convolutional Neural Networks with Test-Time Augmentation
    Wang, Guotai
    Li, Wenqi
    Ourselin, Sebastien
    Vercauteren, Tom
    BRAINLESION: GLIOMA, MULTIPLE SCLEROSIS, STROKE AND TRAUMATIC BRAIN INJURIES, BRAINLES 2018, PT II, 2019, 11384 : 61 - 72
  • [29] Rapid Network Adaptation: Learning to Adapt Neural Networks Using Test-Time Feedback
    Yeo, Teresa
    Kar, Oguzhan Fatih
    Sodagar, Zahra
    Zamir, Amir
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 4651 - 4664
  • [30] GRAPHPATCHER: Mitigating Degree Bias for Graph Neural Networks via Test-time Augmentation
    Ju, Mingxuan
    Zhao, Tong
    Yu, Wenhao
    Shah, Neil
    Ye, Yanfang
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,