Instance-Agnostic and Practical Clean Label Backdoor Attack Method for Deep Learning Based Face Recognition Models

被引:0
|
作者
Kim, Tae-Hoon [1 ]
Choi, Seok-Hwan [2 ]
Choi, Yoon-Ho [1 ]
机构
[1] Pusan Natl Univ, Sch Comp Sci & Engn, Busan 46241, South Korea
[2] Yonsei Univ, Div Software, Wonju 26493, South Korea
基金
新加坡国家研究基金会;
关键词
Face recognition; Training; Labeling; Deep learning; Steganography; Inspection; Filtering; Artificial neural networks; Data poisoning attack; backdoor attack; deep neural networks (DNNs); security;
D O I
10.1109/ACCESS.2023.3342922
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Backdoor attacks, which induce a trained model to behave as intended by an adversary for specific inputs, have recently emerged as a serious security threat in deep learning-based classification models. In particular, because a backdoor attack is executed solely by incorporating a small quantity of malicious data into a dataset, it poses a significant threat to authentication models, such as facial cognition systems. Depending on whether the label of the poisoned samples has been changed, backdoor attacks on deep learning-based face recognition methods are categorized into one of the two architectures: 1) corrupted label attack; and 2) clean label attack. Clean label attack methods have been actively studied because they can be performed without access to training datasets or training processes. However, the performance of previous clean label attack methods is limited in their application to deep learning-based face recognition methods because they only consider digital triggers with instance-specific characteristics. In this study, we propose a novel clean label backdoor attack, that solves the limitations of the scalability of previous clean label attack methods for deep learning-based face recognition models. To generate poisoned samples that are instance agnostic while including physical triggers, the proposed method applies three core techniques: 1) accessory injection; 2) optimization-based feature transfer; and 3) $N$ :1 mapping for generalization. From the experimental results under various conditions, we demonstrate that the proposed attack method is effective for deep learning-based face recognition models in terms of the attack success rate on unseen samples. We also show that the proposed method not only outperforms the recent clean label attack methods, but also maintains a comparable level of classification accuracy when applied to benign data.
引用
收藏
页码:144040 / 144050
页数:11
相关论文
共 22 条
  • [21] New image dataset and new negative sample judgment method for crop pest recognition based on deep learning models
    Wang, Kaili
    Chen, Keyu
    Du, Huiyu
    Liu, Shuang
    Xu, Jingwen
    Zhao, Junfang
    Chen, Houlin
    Liu, Yujun
    Liu, Yang
    ECOLOGICAL INFORMATICS, 2022, 69
  • [22] Defensive Distillation-Based Adversarial Attack Mitigation Method for Channel Estimation Using Deep Learning Models in Next-Generation Wireless Networks
    Catak, Ferhat Ozgur
    Kuzlu, Murat
    Catak, Evren
    Cali, Umit
    Guler, Ozgur
    IEEE ACCESS, 2022, 10 : 98191 - 98203