Backdoor Attack with Imperceptible Input and Latent Modification

被引:0
|
作者
Khoa Doan [1 ]
Lao, Yingjie [1 ]
Li, Ping [1 ]
机构
[1] Baidu Res, Cognit Comp Lab, 10900 NE 8th St, Bellevue, WA 98004 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent studies have shown that deep neural networks (DNN) are vulnerable to various adversarial attacks. In particular, an adversary can inject a stealthy backdoor into a model such that the compromised model will behave normally without the presence of the trigger. Techniques for generating backdoor images that are visually imperceptible from clean images have also been developed recently, which further enhance the stealthiness of the backdoor attacks from the input space. Along with the development of attacks, defense against backdoor attacks is also evolving. Many existing countermeasures found that backdoor tends to leave tangible footprints in the latent or feature space, which can be utilized to mitigate backdoor attacks. In this paper, we extend the concept of imperceptible backdoor from the input space to the latent representation, which significantly improves the effectiveness against the existing defense mechanisms, especially those relying on the distinguishability between clean inputs and backdoor inputs in latent space. In the proposed framework, the trigger function will learn to manipulate the input by injecting imperceptible input noise while matching the latent representations of the clean and manipulated inputs via aWasserstein-based regularization of the corresponding empirical distributions. We formulate such an objective as a non-convex and constrained optimization problem and solve the problem with an efficient stochastic alternating optimization procedure. We name the proposed backdoor attack as Wasserstein Backdoor (WB), which achieves a high attack success rate while being stealthy from both the input and latent spaces, as tested in several benchmark datasets, including MNIST, CIFAR10, GTSRB, and TinyImagenet.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Imperceptible and multi-channel backdoor attack
    Xue, Mingfu
    Ni, Shifeng
    Wu, Yinghao
    Zhang, Yushu
    Liu, Weiqiang
    [J]. APPLIED INTELLIGENCE, 2024, 54 (01) : 1099 - 1116
  • [2] Imperceptible and multi-channel backdoor attack
    Mingfu Xue
    Shifeng Ni
    Yinghao Wu
    Yushu Zhang
    Weiqiang Liu
    [J]. Applied Intelligence, 2024, 54 : 1099 - 1116
  • [3] Imperceptible and Robust Backdoor Attack in 3D Point Cloud
    Gao, Kuofeng
    Bai, Jiawang
    Wu, Baoyuan
    Ya, Mengxi
    Xia, Shu-Tao
    [J]. IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 1267 - 1282
  • [4] RIBAC: Towards Robust and Imperceptible Backdoor Attack against Compact DNN
    Phan, Huy
    Shi, Cong
    Xie, Yi
    Zhang, Tianfang
    Li, Zhuohang
    Zhao, Tianming
    Liu, Jian
    Wang, Yan
    Chen, Yingying
    Yuan, Bo
    [J]. COMPUTER VISION - ECCV 2022, PT IV, 2022, 13664 : 708 - 724
  • [5] Untargeted Backdoor Attack Against Deep Neural Networks With Imperceptible Trigger
    Xue, Mingfu
    Wu, Yinghao
    Ni, Shifeng
    Zhang, Leo Yu
    Zhang, Yushu
    Liu, Weiqiang
    [J]. IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2024, 20 (03) : 5004 - 5013
  • [6] SilentTrig: An imperceptible backdoor attack against speaker identification with hidden triggers
    Tang, Yu
    Sun, Lijuan
    Xu, Xiaolong
    [J]. PATTERN RECOGNITION LETTERS, 2024, 117 : 103 - 109
  • [7] Input-Aware Dynamic Backdoor Attack
    Nguyen, Tuan Anh
    Tran, Tuan Anh
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [8] DEFEAT: Deep Hidden Feature Backdoor Attacks by Imperceptible Perturbation and Latent Representation Constraints
    Zhao, Zhendong
    Chen, Xiaojun
    Xuan, Yuexin
    Dong, Ye
    Wang, Dakui
    Liang, Kaitai
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 15192 - 15201
  • [9] BABE: Backdoor attack with bokeh effects via latent separation suppression
    Li, Junjian
    Chen, Honglong
    Gao, Yudong
    Guo, Shaozhong
    Lin, Kai
    Liu, Yuping
    Sun, Peng
    [J]. Engineering Applications of Artificial Intelligence, 2024, 138
  • [10] An Imperceptible Data Augmentation Based Blackbox Clean-Label Backdoor Attack on Deep Neural Networks
    Xu, Chaohui
    Liu, Wenye
    Zheng, Yue
    Wang, Si
    Chang, Chip-Hong
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2023, 70 (12) : 5011 - 5024