Dual adversarial attacks: Fooling humans and classifiers

被引:2
|
作者
Schneider, Johannes [1 ]
Apruzzese, Giovanni [1 ]
机构
[1] Univ Liechtenstein, Inst Informat Syst, Vaduz, Liechtenstein
关键词
Adversarial attacks; Dual attacks; Computer vision; Deep learning;
D O I
10.1016/j.jisa.2023.103502
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Adversarial samples mostly aim at fooling machine learning (ML) models. They often involve minor pixel-based perturbations that are imperceptible to human observers. In this work, adversarial samples should fool both humans and ML models, which is important in two-stage decision processes. We perform changes on a higher abstraction level so that a target sample exhibits properties of a desired sample. Technically, we contribute by deriving a regularization scheme for autoencoders incorporating a classifier loss for smoothly interpolating between wildly different samples. The realism and effectiveness of generated samples are confirmed with a user study and other evaluations. Our experiments consider neural networks of four architectures, assessed on MNIST, FashionMNIST, QuickDraw and CIFAR-10. Results show that our scheme leads to superior performance compared to existing interpolation techniques: on average, other methods have an 11% higher failure rate when producing a sample that is of any of two interpolated classes. Furthermore, our attacks work in both white -and black-box settings.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Concept-based Adversarial Attacks: Tricking Humans and Classifiers Alike
    Schneider, Johannes
    Apruzzese, Giovanni
    [J]. 2022 43RD IEEE SYMPOSIUM ON SECURITY AND PRIVACY WORKSHOPS (SPW 2022), 2022, : 66 - 72
  • [2] UNIVERSAL ADVERSARIAL ATTACKS ON TEXT CLASSIFIERS
    Behjati, Melika
    Moosavi-Dezfooli, Seyed-Mohsen
    Baghshah, Mahdieh Soleymani
    Frossard, Pascal
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 7345 - 7349
  • [3] ADVERSARIAL ATTACKS ON COARSE-TO-FINE CLASSIFIERS
    Alkhouri, Ismail R.
    Atia, George K.
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 2855 - 2859
  • [4] Are Generative Classifiers More Robust to Adversarial Attacks?
    Li, Yingzhen
    Bradshaw, John
    Sharma, Yash
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [5] Robustness of Sketched Linear Classifiers to Adversarial Attacks
    Mahadevan, Ananth
    Merchant, Arpit
    Wang, Yanhao
    Mathioudakis, Michael
    [J]. PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 4319 - 4323
  • [6] Adversarial scratches: Deployable attacks to CNN classifiers
    Giulivi, Loris
    Jere, Malhar
    Rossi, Loris
    Koushanfar, Farinaz
    Ciocarlie, Gabriela
    Hitaj, Briland
    Boracchi, Giacomo
    [J]. PATTERN RECOGNITION, 2023, 133
  • [7] Fooling LIME and SHAP: Adversarial Attacks on Post hoc Explanation Methods
    Slack, Dylan
    Hilgard, Sophie
    Jia, Emily
    Singh, Sameer
    Lakkaraju, Himabindu
    [J]. PROCEEDINGS OF THE 3RD AAAI/ACM CONFERENCE ON AI, ETHICS, AND SOCIETY AIES 2020, 2020, : 180 - 186
  • [8] Fooling AI with AI: An Accelerator for Adversarial Attacks on Deep Learning Visual Classification
    Guo, Haoqiang
    Peng, Lu
    Zhang, Jian
    Qi, Fang
    Duan, Lide
    [J]. 2019 IEEE 30TH INTERNATIONAL CONFERENCE ON APPLICATION-SPECIFIC SYSTEMS, ARCHITECTURES AND PROCESSORS (ASAP 2019), 2019, : 136 - 136
  • [9] Evaluation of adversarial attacks sensitivity of classifiers with occluded input data
    Korn Sooksatra
    Pablo Rivas
    [J]. Neural Computing and Applications, 2022, 34 : 17615 - 17632
  • [10] Evaluation of adversarial attacks sensitivity of classifiers with occluded input data
    Sooksatra, Korn
    Rivas, Pablo
    [J]. NEURAL COMPUTING & APPLICATIONS, 2022, 34 (20): : 17615 - 17632