Improving query efficiency of black-box attacks via the preference of models

被引:1
|
作者
Yang, Xiangyuan [1 ]
Lin, Jie [1 ]
Zhang, Hanlin [2 ]
Zhao, Peng [1 ]
机构
[1] Xi'an Jiaotong Univ, Sch Comp Sci & Technol, Xian, Peoples R China
[2] Qingdao Univ, Qingdao, Peoples R China
关键词
Black-box query attack; Gradient-aligned attack; Preference property; Gradient preference; ROBUSTNESS;
D O I
10.1016/j.ins.2024.121013
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Black-box query attacks are effective at compromising deep-learning models using only the model's output. These attacks typically face challenges with low attack success rates (ASRs) when limited to fewer than ten queries per example. Recent approaches have improved ASRs due to the transferability of initial perturbations, yet they still suffer from inefficient querying. Our study introduces the Gradient-Aligned Attack (GAA) to enhance ASRs with minimal perturbation by focusing on the model's preference. We define a preference property where the generated adversarial example prefers to be misclassified as the wrong category with a high initial confidence. This property is further elucidated by the gradient preference, suggesting a positive correlation between the magnitude of a coefficient in a partial derivative and the norm of the derivative itself. Utilizing this, we devise the gradient-aligned CE (GACE) loss to precisely estimate gradients by aligning these coefficients between the surrogate and victim models, with coefficients assessed by the victim model's outputs. GAA, based on the GACE loss, also aims to achieve the smallest perturbation. Our tests on ImageNet, CIFAR10, and Imagga API show that GAA can increase ASRs by 25.7% and 40.3% for untargeted and targeted attacks respectively, while only needing minimally disruptive perturbations. Furthermore, the GACE loss reduces the number of necessary queries by up to 2.5x and enhances the transferability of advanced attacks by up to 14.2%, especially when using an ensemble surrogate model. Code is available at https:// github .com /HaloMoto /GradientAlignedAttack.
引用
收藏
页数:21
相关论文
共 50 条
  • [21] Efficient Label Contamination Attacks Against Black-Box Learning Models
    Zhao, Mengchen
    An, Bo
    Gao, Wei
    Zhang, Teng
    PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 3945 - 3951
  • [22] Black-box adversarial attacks against image quality assessment models
    Ran, Yu
    Zhang, Ao-Xiang
    Li, Mingjie
    Tang, Weixuan
    Wang, Yuan-Gen
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 260
  • [23] Improving the expressiveness of black-box models for predicting student performance
    Villagra-Arnedo, Carlos J.
    Gallego-Duran, Francisco J.
    Llorens-Largo, Faraon
    Compan-Rosique, Patricia
    Satorre-Cuerda, Rosana
    Molina-Carmona, Rafael
    COMPUTERS IN HUMAN BEHAVIOR, 2017, 72 : 621 - 631
  • [24] Mitigating Black-Box Adversarial Attacks via Output Noise Perturbation
    Aithal, Manjushree B.
    Li, Xiaohua
    IEEE ACCESS, 2022, 10 : 12395 - 12411
  • [25] Automatic Selection Attacks Framework for Hard Label Black-Box Models
    Liu, Xiaolei
    Li, Xiaoyu
    Zheng, Desheng
    Bai, Jiayu
    Peng, Yu
    Zhang, Shibin
    IEEE INFOCOM 2022 - IEEE CONFERENCE ON COMPUTER COMMUNICATIONS WORKSHOPS (INFOCOM WKSHPS), 2022,
  • [26] Black-box attacks on dynamic graphs via adversarial topology perturbations
    Tao, Haicheng
    Cao, Jie
    Chen, Lei
    Sun, Hongliang
    Shi, Yong
    Zhu, Xingquan
    NEURAL NETWORKS, 2024, 171 : 308 - 319
  • [27] THE BLACK-BOX QUERY COMPLEXITY OF POLYNOMIAL SUMMATION
    Juma, Ali
    Kabanets, Valentine
    Rackoff, Charles
    Shpilka, Amir
    COMPUTATIONAL COMPLEXITY, 2009, 18 (01) : 59 - 79
  • [28] The Black-Box Query Complexity of Polynomial Summation
    Ali Juma
    Valentine Kabanets
    Charles Rackoff
    Amir Shpilka
    computational complexity, 2009, 18 : 59 - 79
  • [29] Parsimonious Black-Box Adversarial Attacks via Efficient Combinatorial Optimization
    Moon, Seungyong
    An, Gaon
    Song, Hyun Oh
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [30] Enhancing Transferability of Black-box Adversarial Attacks via Lifelong Learning for Speech Emotion Recognition Models
    Ren, Zhao
    Han, Jing
    Cummins, Nicholas
    Schuller, Bjoern W.
    INTERSPEECH 2020, 2020, : 496 - 500