Sparse Backdoor Attack Against Neural Networks

被引:0
|
作者
Zhong, Nan [1 ]
Qian, Zhenxing [1 ]
Zhang, Xinpeng [1 ]
机构
[1] Fudan Univ, Sch Comp Sci, Shanghai 200438, Peoples R China
来源
COMPUTER JOURNAL | 2023年 / 67卷 / 05期
基金
中国国家自然科学基金;
关键词
Backdoor attack; AI security; Trustworthy AI;
D O I
10.1093/comjnl/bxad100
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Recent studies show that neural networks are vulnerable to backdoor attacks, in which compromised networks behave normally for clean inputs but make mistakes when a pre-defined trigger appears. Although prior studies have designed various invisible triggers to avoid causing visual anomalies, they cannot evade some trigger detectors. In this paper, we consider the stealthiness of backdoor attacks from input space and feature representation space. We propose a novel backdoor attack named sparse backdoor attack, and investigate the minimum required trigger to induce the well-trained networks to make incorrect results. A U-net-based generator is employed to create triggers for each clean image. Considering the stealthiness of the trigger, we restrict the elements of the trigger between -1 and 1. In the aspect of the feature representation domain, we adopt an entanglement cost function to minimize the gap between feature representations of benign and malicious inputs. The inseparability of benign and malicious feature representations contributes to the stealthiness of our attack against various model diagnosis-based defences. We validate the effectiveness and generalization of our method by conducting extensive experiments on multiple datasets and networks.
引用
收藏
页码:1783 / 1793
页数:11
相关论文
共 50 条
  • [1] Adaptive Backdoor Attack against Deep Neural Networks
    He, Honglu
    Zhu, Zhiying
    Zhang, Xinpeng
    [J]. CMES-COMPUTER MODELING IN ENGINEERING & SCIENCES, 2023, 136 (03): : 2617 - 2633
  • [2] A backdoor attack against quantum neural networks with limited information
    Huang, Chen-Yi
    Zhang, Shi-Bin
    [J]. CHINESE PHYSICS B, 2023, 32 (10)
  • [3] A backdoor attack against quantum neural networks with limited information
    黄晨猗
    张仕斌
    [J]. Chinese Physics B, 2023, 32 (10) : 260 - 269
  • [4] SGBA: A stealthy scapegoat backdoor attack against deep neural networks
    He, Ying
    Shen, Zhili
    Xia, Chang
    Hua, Jingyu
    Tong, Wei
    Zhong, Sheng
    [J]. COMPUTERS & SECURITY, 2024, 136
  • [5] Compression-resistant backdoor attack against deep neural networks
    Mingfu Xue
    Xin Wang
    Shichang Sun
    Yushu Zhang
    Jian Wang
    Weiqiang Liu
    [J]. Applied Intelligence, 2023, 53 : 20402 - 20417
  • [6] Stealthy dynamic backdoor attack against neural networks for image classification
    Dong, Liang
    Qiu, Jiawei
    Fu, Zhongwang
    Chen, Leiyang
    Cui, Xiaohui
    Shen, Zhidong
    [J]. APPLIED SOFT COMPUTING, 2023, 149
  • [7] Compression-resistant backdoor attack against deep neural networks
    Xue, Mingfu
    Wang, Xin
    Sun, Shichang
    Zhang, Yushu
    Wang, Jian
    Liu, Weiqiang
    [J]. APPLIED INTELLIGENCE, 2023, 53 (17) : 20402 - 20417
  • [8] Untargeted Backdoor Attack Against Deep Neural Networks With Imperceptible Trigger
    Xue, Mingfu
    Wu, Yinghao
    Ni, Shifeng
    Zhang, Leo Yu
    Zhang, Yushu
    Liu, Weiqiang
    [J]. IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2024, 20 (03) : 5004 - 5013
  • [9] PatchBackdoor: Backdoor Attack against Deep Neural Networks without Model Modification
    Yuan, Yizhen
    Kong, Rui
    Xie, Shenghao
    Li, Yuanchun
    Liu, Yunxin
    [J]. PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 9134 - 9142
  • [10] BadRL: Sparse Targeted Backdoor Attack against Reinforcement Learning
    Cui, Jing
    Han, Yufei
    Ma, Yuzhe
    Jiao, Jianbin
    Zhang, Junge
    [J]. THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 10, 2024, : 11687 - 11694