Untargeted Backdoor Attack Against Deep Neural Networks With Imperceptible Trigger

被引:0
|
作者
Xue, Mingfu [1 ]
Wu, Yinghao [1 ]
Ni, Shifeng [1 ]
Zhang, Leo Yu [2 ]
Zhang, Yushu [1 ]
Liu, Weiqiang [3 ]
机构
[1] Nanjing Univ Aeronaut & Astronaut, Coll Comp Sci & Technol, Nanjing 211106, Peoples R China
[2] Griffith Univ, Sch Informat & Commun Technol, Nathan, Qld 4111, Australia
[3] Nanjing Univ Aeronaut & Astronaut, Coll Elect & Informat Engn, Nanjing 211106, Peoples R China
基金
中国国家自然科学基金;
关键词
Training; Predictive models; Artificial neural networks; Entropy; Aerospace electronics; Informatics; Force; Autoencoder; deep neural networks (DNNs); imperceptible trigger; trustworthy artificial intelligence; untargeted backdoor attack (UBA);
D O I
10.1109/TII.2023.3329641
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recent research works have demonstrated that deep neural networks (DNNs) are vulnerable to backdoor attacks. The existing backdoor attacks can only cause targeted misclassification on backdoor instances, which makes them can be easily detected by defense methods. In this article, we propose an untargeted backdoor attack (UBA) against DNNs, where the backdoor instances are randomly misclassified by the backdoored model to any incorrect label. To achieve the goal of UBA, we propose to utilize autoencoder as the trigger generation model and train the target model and the autoencoder simultaneously. We also propose a special loss function (Evasion Loss) to train the autoencoder and the target model, in order to make the target model predict backdoor instances as random incorrect classes. During the inference stage, the trained autoencoder is used to generate backdoor instances. For different backdoor instances, the generated triggers are different and the corresponding predicted labels are random incorrect labels. Experimental results demonstrate that the proposed UBA is effective. On the ResNet-18 model, the attack success rate (ASR) of the proposed UBA is 96.48%, 91.27%, and 90.83% on CIFAR-10, GTSRB, and ImageNet datasets, respectively. On the VGG-16 model, the ASR of the proposed UBA is 89.72% and 97.78% on CIFAR-10 and ImageNet datasets, respectively. Moreover, the proposed UBA is robust against existing backdoor defense methods, which are designed to detect targeted backdoor attacks. We hope this article can promote the research of corresponding backdoor defense works.
引用
收藏
页码:5004 / 5013
页数:10
相关论文
共 50 条
  • [31] Imperceptible graph injection attack on graph neural networks
    Yang Chen
    Zhonglin Ye
    Zhaoyang Wang
    Haixing Zhao
    [J]. Complex & Intelligent Systems, 2024, 10 : 869 - 883
  • [32] Backdoor smoothing: Demystifying backdoor attacks on deep neural networks
    Grosse, Kathrin
    Lee, Taesung
    Biggio, Battista
    Park, Youngja
    Backes, Michael
    Molloy, Ian
    [J]. Computers and Security, 2022, 120
  • [33] Backdoor smoothing: Demystifying backdoor attacks on deep neural networks
    Grosse, Kathrin
    Lee, Taesung
    Biggio, Battista
    Park, Youngja
    Backes, Michael
    Molloy, Ian
    [J]. COMPUTERS & SECURITY, 2022, 120
  • [34] Imperceptible Adversarial Attack via Invertible Neural Networks
    Chen, Zihan
    Wang, Ziyue
    Huang, Jun-Jie
    Zhao, Wentao
    Liu, Xiao
    Guan, Dejian
    [J]. THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 1, 2023, : 414 - 424
  • [36] Backdoor Attack With Sparse and Invisible Trigger
    Gao, Yinghua
    Li, Yiming
    Gong, Xueluan
    Li, Zhifeng
    Xia, Shu-Tao
    Wang, Qian
    [J]. IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 6364 - 6376
  • [37] A multitarget backdooring attack on deep neural networks with random location trigger
    Xiao, Yu
    Cong, Liu
    Mingwen, Zheng
    Yajie, Wang
    Xinrui, Liu
    Shuxiao, Song
    Yuexuan, Ma
    Jun, Zheng
    [J]. INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2022, 37 (03) : 2567 - 2583
  • [38] A semantic backdoor attack against graph convolutional networks
    Dai, Jiazhu
    Xiong, Zhipeng
    Cao, Chenhong
    [J]. NEUROCOMPUTING, 2024, 600
  • [39] Diffense: Defense Against Backdoor Attacks on Deep Neural Networks With Latent Diffusion
    Hu, Bowen
    Chang, Chip-Hong
    [J]. IEEE Journal on Emerging and Selected Topics in Circuits and Systems, 2024, 14 (04) : 729 - 742
  • [40] Interpretability-Guided Defense Against Backdoor Attacks to Deep Neural Networks
    Jiang, Wei
    Wen, Xiangyu
    Zhan, Jinyu
    Wang, Xupeng
    Song, Ziwei
    [J]. IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2022, 41 (08) : 2611 - 2624