Compression-resistant backdoor attack against deep neural networks

被引:2
|
作者
Xue, Mingfu [1 ]
Wang, Xin [1 ]
Sun, Shichang [1 ]
Zhang, Yushu [1 ]
Wang, Jian [1 ]
Liu, Weiqiang [2 ]
机构
[1] Nanjing Univ Aeronaut & Astronaut, Coll Comp Sci & Technol, Nanjing, Peoples R China
[2] Nanjing Univ Aeronaut & Astronaut, Coll Elect & Informat Engn, Nanjing, Peoples R China
基金
中国国家自然科学基金;
关键词
Artificial intelligence security; Backdoor attack; Compression resistance; Deep neural networks; Feature consistency training;
D O I
10.1007/s10489-023-04575-8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In recent years, a number of backdoor attacks against deep neural networks (DNN) have been proposed. In this paper, we reveal that backdoor attacks are vulnerable to image compressions, as backdoor instances used to trigger backdoor attacks are usually compressed by image compression methods during data transmission. When backdoor instances are compressed, the feature of backdoor trigger will be destroyed, which could result in significant performance degradation for backdoor attacks. As a countermeasure, we propose the first compression-resistant backdoor attack method based on feature consistency training. Specifically, both backdoor images and their compressed versions are used for training, and the feature difference between backdoor images and their compressed versions are minimized through feature consistency training. As a result, the DNN treats the feature of compressed images as the feature of backdoor images in feature space. After training, the backdoor attack will be robust to image compressions. Furthermore, we consider three different image compressions (i.e., JPEG, JPEG2000, WEBP) during the feature consistency training, so that the backdoor attack can be robust to multiple image compression algorithms. Experimental results demonstrate that when the backdoor instances are compressed, the attack success rate of common backdoor attack is 6.63% (JPEG), 6.20% (JPEG2000) and 3.97% (WEBP) respectively, while the attack success rate of the proposed compression-resistant backdoor attack is 98.77% (JPEG), 97.69% (JPEG2000), and 98.93% (WEBP) respectively. The compression-resistant attack is robust under various parameters settings. In addition, extensive experiments have demonstrated that even if only one image compression method is used in the feature consistency training process, the proposed compression-resistant backdoor attack has the generalization ability to resist multiple unseen image compression methods.
引用
收藏
页码:20402 / 20417
页数:16
相关论文
共 50 条
  • [21] An Overview of Backdoor Attacks Against Deep Neural Networks and Possible Defences
    Guo, Wei
    Tondi, Benedetta
    Barni, Mauro
    [J]. IEEE OPEN JOURNAL OF SIGNAL PROCESSING, 2022, 3 : 261 - 287
  • [22] Backdoor Attacks against Deep Neural Networks by Personalized Audio Steganography
    Liu, Peng
    Zhang, Shuyi
    Yao, Chuanjian
    Ye, Wenzhe
    Li, Xianxian
    [J]. 2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 68 - 74
  • [23] A Backdoor Embedding Method for Backdoor Detection in Deep Neural Networks
    Liu, Meirong
    Zheng, Hong
    Liu, Qin
    Xing, Xiaofei
    Dai, Yinglong
    [J]. UBIQUITOUS SECURITY, 2022, 1557 : 1 - 12
  • [24] Backdoor smoothing: Demystifying backdoor attacks on deep neural networks
    Grosse, Kathrin
    Lee, Taesung
    Biggio, Battista
    Park, Youngja
    Backes, Michael
    Molloy, Ian
    [J]. COMPUTERS & SECURITY, 2022, 120
  • [25] Backdoor smoothing: Demystifying backdoor attacks on deep neural networks
    Grosse, Kathrin
    Lee, Taesung
    Biggio, Battista
    Park, Youngja
    Backes, Michael
    Molloy, Ian
    [J]. Computers and Security, 2022, 120
  • [27] Diffense: Defense Against Backdoor Attacks on Deep Neural Networks With Latent Diffusion
    Hu, Bowen
    Chang, Chip-Hong
    [J]. IEEE Journal on Emerging and Selected Topics in Circuits and Systems, 2024, 14 (04) : 729 - 742
  • [28] Deep Model Intellectual Property Protection with Compression-Resistant Model Watermarking
    Nie H.
    Lu S.
    Wu J.
    Zhu J.
    [J]. IEEE Transactions on Artificial Intelligence, 2024, 5 (07): : 1 - 12
  • [29] A semantic backdoor attack against graph convolutional networks
    Dai, Jiazhu
    Xiong, Zhipeng
    Cao, Chenhong
    [J]. NEUROCOMPUTING, 2024, 600
  • [30] Latent Space-Based Backdoor Attacks Against Deep Neural Networks
    Kristanto, Adrian
    Wang, Shuo
    Rudolph, Carsten
    [J]. 2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,