Mitigating Black-Box Adversarial Attacks via Output Noise Perturbation

被引:1
|
作者
Aithal, Manjushree B. [1 ]
Li, Xiaohua [1 ]
机构
[1] Binghamton Univ, Dept Elect & Comp Engn, Binghamton, NY 13902 USA
关键词
Perturbation methods; Signal to noise ratio; Standards; Noise level; White noise; Noise measurement; Neural networks; Deep learning; adversarial machine learning; black-box attack; noise perturbation; performance analysis;
D O I
10.1109/ACCESS.2022.3146198
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In black-box adversarial attacks, attackers query the deep neural network (DNN) and use the query results to optimize the adversarial samples iteratively. In this paper, we study the method of adding white noise to the DNN output to mitigate such attacks. One of our unique contributions is a theoretical analysis of gradient signal-to-noise ratio (SNR), which shows the trade-off between the defense noise level and the attack query cost. The attacker's query count (QC) is derived mathematically as a function of noise standard deviation. This will guide the defender to find the appropriate noise level for mitigating attacks to the desired security level specified by QC and DNN performance loss. Our analysis shows that the added noise is drastically magnified by the small variation of DNN outputs, which makes the reconstructed gradient have an extremely low SNR. Adding slight white noise with a very small standard deviation, e.g., less than 0.01, is enough to increase QC by many orders of magnitude yet without introducing any noticeable classification accuracy reduction. Our experiments demonstrate that this method can effectively mitigate both soft-label and hard-label black-box attacks under realistic QC constraints. We also prove that this method outperforms many other defense methods and is robust to the attacker's countermeasures.
引用
收藏
页码:12395 / 12411
页数:17
相关论文
共 50 条
  • [31] Semantics aware adversarial malware examples generation for black-box attacks
    Peng, Xiaowei
    Xian, Hequn
    Lu, Qian
    Lu, Xiuqing
    APPLIED SOFT COMPUTING, 2021, 109
  • [32] Black-box attacks against log anomaly detection with adversarial examples
    Lu, Siyang
    Wang, Mingquan
    Wang, Dongdong
    Wei, Xiang
    Xiao, Sizhe
    Wang, Zhiwei
    Han, Ningning
    Wang, Liqiang
    INFORMATION SCIENCES, 2023, 619 : 249 - 262
  • [33] Efficient Local Imperceptible Random Search for Black-Box Adversarial Attacks
    Li, Yining
    You, Shu
    Chen, Yihan
    Li, Zhenhua
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT XI, ICIC 2024, 2024, 14872 : 325 - 336
  • [34] Improving the transferability of adversarial examples through black-box feature attacks
    Wang, Maoyuan
    Wang, Jinwei
    Ma, Bin
    Luo, Xiangyang
    NEUROCOMPUTING, 2024, 595
  • [35] Black-box Attacks on Spoofing Countermeasures Using Transferability of Adversarial Examples
    Zhang, Yuekai
    Jiang, Ziyan
    Villalba, Jesus
    Dehak, Najim
    INTERSPEECH 2020, 2020, : 4238 - 4242
  • [36] Adversarial Black-Box Attacks with Timing Side-Channel Leakage
    Nakai, Tsunato
    Suzuki, Daisuke
    Omatsu, Fumio
    Fujino, Takeshi
    IEICE TRANSACTIONS ON FUNDAMENTALS OF ELECTRONICS COMMUNICATIONS AND COMPUTER SCIENCES, 2021, E104A (01) : 143 - 151
  • [37] Black-box Adversarial Attacks on Commercial Speech Platforms with Minimal Information
    Zhene, Baolin
    Jiang, Peipei
    Wang, Qian
    Li, Qi
    Shen, Chao
    Wang, Cong
    Ge, Yunjie
    Teng, Qingyang
    Zhang, Shenyi
    CCS '21: PROCEEDINGS OF THE 2021 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2021, : 86 - 107
  • [38] Simultaneously Optimizing Perturbations and Positions for Black-Box Adversarial Patch Attacks
    Wei, Xingxing
    Guo, Ying
    Yu, Jie
    Zhang, Bo
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (07) : 9041 - 9054
  • [39] Data-free Universal Adversarial Perturbation and Black-box Attack
    Zhang, Chaoning
    Benz, Philipp
    Karjauv, Adil
    Kweon, In So
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 7848 - 7857
  • [40] Improving Black-box Adversarial Attacks with a Transfer-based Prior
    Cheng, Shuyu
    Dong, Yinpeng
    Pang, Tianyu
    Su, Hang
    Zhu, Jun
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32