Adversarial Robustness of Model Sets

被引:1
|
作者
Megyeri, Istvan [1 ]
Hegedus, Istvan [1 ]
Jelasity, Mark [1 ,2 ]
机构
[1] Univ Szeged, Szeged, Hungary
[2] MTA SZTE Res Grp AI, Szeged, Hungary
关键词
adversarial examples; multi-model attack; model set attack; DeepFool; deep neural networks;
D O I
10.1109/ijcnn48605.2020.9206656
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Machine learning models are vulnerable to very small adversarial input perturbations. Here, we study the question of whether the list of predictions made by a list of models can also be changed arbitrarily by a single small perturbation. Clearly, this is a harder problem since one has to simultaneously mislead several models using the same perturbation, where the target classes assigned to the models might differ. This attack has several applications over models designed by different manufacturers for a similar purpose. One might want a single perturbation that acts differently on each model; like only misleading a subset, or making each model predict a different label. Also, one might want a perturbation that misleads each model the same way and thereby create a transferable perturbation. Current approaches are not applicable for this general problem directly. Here, we propose an algorithm that is able to find a perturbation that satisfies several kinds of attack patterns. For example, all the models could have the same target class, or different random target classes, or target classes designed to be maximally contradicting. We evaluated our algorithm using three model sets consisting of publicly available pre-trained ImageNet models of varying capacity and architecture. We demonstrate that, in all the scenarios, our method is able to find visually insigni ficant perturbations that achieve our target adversarial patterns.(1)
引用
收藏
页数:8
相关论文
共 50 条
  • [1] Reinforcing Adversarial Robustness using Model Confidence Induced by Adversarial Training
    Wu, Xi
    Jang, Uyeong
    Chen, Jiefeng
    Chen, Lingjiao
    Jha, Somesh
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [2] Enhancing Model Robustness Against Adversarial Attacks with an Anti-adversarial Module
    Qin, Zhiquan
    Liu, Guoxing
    Lin, Xianming
    [J]. PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT IX, 2024, 14433 : 66 - 78
  • [3] Achieving Model Robustness through Discrete Adversarial Training
    Ivgi, Maor
    Berant, Jonathan
    [J]. 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 1529 - 1544
  • [4] Model Compression with Adversarial Robustness: A Unified Optimization Framework
    Gui, Shupeng
    Wang, Haotao
    Yang, Haichuan
    Yu, Chen
    Wang, Zhangyang
    Liu, Ji
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [5] Adversarial Robustness vs. Model Compression, or Both?
    Ye, Shaokai
    Xu, Kaidi
    Liu, Sijia
    Cheng, Hao
    Lambrechts, Jan-Henrik
    Zhang, Huan
    Zhou, Aojun
    Ma, Kaisheng
    Wang, Yanzhi
    Lin, Xue
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 111 - 120
  • [6] Improving Adversarial Robustness With Adversarial Augmentations
    Chen, Chuanxi
    Ye, Dengpan
    He, Yiheng
    Tang, Long
    Xu, Yue
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (03) : 5105 - 5117
  • [7] Adversarial Robustness for Code
    Bielik, Pavol
    Vechev, Martin
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [8] Adversarial Robustness Curves
    Goepfert, Christina
    Goepfert, Jan Philip
    Hammer, Barbara
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2019, PT I, 2020, 1167 : 172 - 179
  • [9] The Adversarial Robustness of Sampling
    Ben-Eliezer, Omri
    Yogev, Eylon
    [J]. PODS'20: PROCEEDINGS OF THE 39TH ACM SIGMOD-SIGACT-SIGAI SYMPOSIUM ON PRINCIPLES OF DATABASE SYSTEMS, 2020, : 49 - 62
  • [10] Incrementing Adversarial Robustness with Autoencoding for Machine Learning Model Attacks
    Sivaslioglu, Salved
    Catak, Ferhat Ozgur
    Gul, Ensar
    [J]. 2019 27TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2019,