Mitigating Black-Box Adversarial Attacks via Output Noise Perturbation

被引:1
|
作者
Aithal, Manjushree B. [1 ]
Li, Xiaohua [1 ]
机构
[1] Binghamton Univ, Dept Elect & Comp Engn, Binghamton, NY 13902 USA
关键词
Perturbation methods; Signal to noise ratio; Standards; Noise level; White noise; Noise measurement; Neural networks; Deep learning; adversarial machine learning; black-box attack; noise perturbation; performance analysis;
D O I
10.1109/ACCESS.2022.3146198
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In black-box adversarial attacks, attackers query the deep neural network (DNN) and use the query results to optimize the adversarial samples iteratively. In this paper, we study the method of adding white noise to the DNN output to mitigate such attacks. One of our unique contributions is a theoretical analysis of gradient signal-to-noise ratio (SNR), which shows the trade-off between the defense noise level and the attack query cost. The attacker's query count (QC) is derived mathematically as a function of noise standard deviation. This will guide the defender to find the appropriate noise level for mitigating attacks to the desired security level specified by QC and DNN performance loss. Our analysis shows that the added noise is drastically magnified by the small variation of DNN outputs, which makes the reconstructed gradient have an extremely low SNR. Adding slight white noise with a very small standard deviation, e.g., less than 0.01, is enough to increase QC by many orders of magnitude yet without introducing any noticeable classification accuracy reduction. Our experiments demonstrate that this method can effectively mitigate both soft-label and hard-label black-box attacks under realistic QC constraints. We also prove that this method outperforms many other defense methods and is robust to the attacker's countermeasures.
引用
收藏
页码:12395 / 12411
页数:17
相关论文
共 50 条
  • [21] Black-box transferable adversarial attacks based on ensemble advGAN
    Huang S.-N.
    Li Y.-X.
    Mao Y.-H.
    Ban A.-Y.
    Zhang Z.-Y.
    Jilin Daxue Xuebao (Gongxueban)/Journal of Jilin University (Engineering and Technology Edition), 2022, 52 (10): : 2391 - 2398
  • [22] Black-Box Adversarial Attacks against Audio Forensics Models
    Jiang, Yi
    Ye, Dengpan
    SECURITY AND COMMUNICATION NETWORKS, 2022, 2022
  • [23] AutoAttacker: A reinforcement learning approach for black-box adversarial attacks
    Tsingenopoulos, Ilias
    Preuveneers, Davy
    Joosen, Wouter
    2019 4TH IEEE EUROPEAN SYMPOSIUM ON SECURITY AND PRIVACY WORKSHOPS (EUROS&PW), 2019, : 229 - 237
  • [24] Query-based Local Black-box Adversarial Attacks
    Shi, Jing
    Zhang, Xiaolin
    Xu, Enhui
    Wang, Yongping
    Zhang, Wenwen
    International Journal of Network Security, 2023, 25 (06) : 1048 - 1058
  • [25] Simple Black-Box Adversarial Attacks on Deep Neural Networks
    Narodytska, Nina
    Kasiviswanathan, Shiva
    2017 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2017, : 1310 - 1318
  • [26] Heuristic Black-Box Adversarial Attacks on Video Recognition Models
    Wei, Zhipeng
    Chen, Jingjing
    Wei, Xingxing
    Jiang, Linxi
    Chua, Tat-Seng
    Zhou, Fengfeng
    Jiang, Yu-Gang
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 12338 - 12345
  • [27] Sensitive region-aware black-box adversarial attacks
    Lin, Chenhao
    Han, Sicong
    Zhu, Jiongli
    Li, Qian
    Shen, Chao
    Zhang, Youwei
    Guan, Xiaohong
    INFORMATION SCIENCES, 2023, 637
  • [28] Adaptive Temporal Grouping for Black-box Adversarial Attacks on Videos
    Wei, Zhipeng
    Chen, Jingjing
    Zhang, Hao
    Jiang, Linxi
    Jiang, Yu-Gang
    PROCEEDINGS OF THE 2022 INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2022, 2022, : 587 - 593
  • [29] Black-Box Adversarial Attack via Overlapped Shapes
    Williams, Phoenix
    Li, Ke
    Min, Geyong
    PROCEEDINGS OF THE 2022 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE COMPANION, GECCO 2022, 2022, : 467 - 468
  • [30] HyGloadAttack: Hard-label black-box textual adversarial attacks via hybrid optimization
    Liu, Zhaorong
    Xiong, Xi
    Li, Yuanyuan
    Yu, Yan
    Lu, Jiazhong
    Zhang, Shuai
    Xiong, Fei
    NEURAL NETWORKS, 2024, 178