TEST-TIME DETECTION OF BACKDOOR TRIGGERS FOR POISONED DEEP NEURAL NETWORKS

被引:4
|
作者
Li, Xi [1 ]
Xiang, Zhen [1 ]
Miller, David J. [1 ]
Kesidis, George [1 ]
机构
[1] Penn State Univ, Sch EECS, Philadelphia, PA 19104 USA
关键词
adversarial learning; backdoor attack; Trojan attack; in-flight detection; image classification;
D O I
10.1109/ICASSP43922.2022.9746573
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Backdoor (Trojan) attacks are emerging threats against deep neural networks (DNN). A DNN being attacked will predict to an attacker-desired target class whenever a test sample from any source class is embedded with a backdoor pattern, while correctly classifying clean (attack-free) test samples. Existing backdoor defenses have shown success in detecting whether a DNN is attacked and in reverse-engineering the backdoor pattern in a "post-training" scenario: the defender has access to the DNN to be inspected and a small, clean dataset collected independently, but has no access to the (possibly poisoned) training set of the DNN. However, these defenses neither catch culprits in the act of triggering the backdoor mapping, nor mitigate the backdoor attack at testtime. In this paper, we propose an "in-flight" unsupervised defense against backdoor attacks on image classification that 1) detects use of a backdoor trigger at test-time; and 2) infers the class of origin (source class) for a detected trigger example. The effectiveness of our defense is demonstrated experimentally for a wide variety of DNN architectures, datasets, and backdoor attack configurations.
引用
收藏
页码:3333 / 3337
页数:5
相关论文
共 50 条
  • [1] TEST-TIME DETECTION OF BACKDOOR TRIGGERS FOR POISONED DEEP NEURAL NETWORKS
    Li, Xi
    Xiang, Zhen
    Miller, David J.
    Kesidis, George
    ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings, 2022, 2022-May : 3333 - 3337
  • [2] Test-time detection of backdoor triggers for poisoned deep neural networks
    Li, Xi
    Xiang, Zhen
    Miller, David J.
    Kesidis, George
    arXiv, 2021,
  • [3] A Backdoor Embedding Method for Backdoor Detection in Deep Neural Networks
    Liu, Meirong
    Zheng, Hong
    Liu, Qin
    Xing, Xiaofei
    Dai, Yinglong
    UBIQUITOUS SECURITY, 2022, 1557 : 1 - 12
  • [4] IMPROVED ACTIVATION CLIPPING FOR UNIVERSAL BACKDOOR MITIGATION AND TEST-TIME DETECTION
    Anomalee Inc. & Pennsylvania State University, United States
    不详
    arXiv,
  • [5] Reliable Prediction Errors for Deep Neural Networks Using Test-Time Dropout
    Cortes-Ciriano, Isidro
    Bender, Andreas
    JOURNAL OF CHEMICAL INFORMATION AND MODELING, 2019, 59 (07) : 3330 - 3339
  • [6] Unsupervised Test-Time Adaptation of Deep Neural Networks at the Edge: A Case Study
    Bhardwaj, Kshitij
    Diffenderfer, James
    Kailkhura, Bhavya
    Gokhale, Maya
    PROCEEDINGS OF THE 2022 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE 2022), 2022, : 412 - 417
  • [7] Backdoor Defense via Test-Time Detecting and Repairing
    Guan, Jiyang
    Liang, Jian
    He, Ran
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 24564 - 24573
  • [8] Universal backdoor attack on deep neural networks for malware detection
    Zhang, Yunchun
    Feng, Fan
    Liao, Zikun
    Li, Zixuan
    Yao, Shaowen
    APPLIED SOFT COMPUTING, 2023, 143
  • [9] An Approach to Generation Triggers for Parrying Backdoor in Neural Networks
    Artem, Menisov
    ARTIFICIAL GENERAL INTELLIGENCE, AGI 2022, 2023, 13539 : 304 - 314
  • [10] Transfer Learning via Test-time Neural Networks Aggregation
    Casella, Bruno
    Chisari, Alessio Barbaro
    Battiato, Sebastiano
    Giuffrida, Mario Valerio
    PROCEEDINGS OF THE 17TH INTERNATIONAL JOINT CONFERENCE ON COMPUTER VISION, IMAGING AND COMPUTER GRAPHICS THEORY AND APPLICATIONS (VISAPP), VOL 5, 2022, : 642 - 649