Aha! Adaptive History-driven Attack for Decision-based Black-box Models

被引:4
|
作者
Li, Jie [1 ]
Ji, Rongrong [1 ,2 ,4 ]
Chen, Peixian [1 ,6 ]
Zhang, Baochang [3 ]
Hong, Xiaopeng [5 ]
Zhang, Ruixin [6 ]
Li, Shaoxin [6 ]
Li, Jilin [6 ]
Huang, Feiyue [6 ]
Wu, Yongjian [6 ]
机构
[1] Xiamen Univ, MAC Lab, Sch Informat, Xiamen, Peoples R China
[2] Peng Cheng Lab, Xiamen, Peoples R China
[3] Beihang Univ, Beijing, Peoples R China
[4] Xiamen Univ, Inst Artificial Intelligence, Xiamen, Peoples R China
[5] Xi An Jiao Tong Univ, Xian, Peoples R China
[6] Tencent, Youtu Lab, Shenzhen, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1109/ICCV48922.2021.01586
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The decision-based black-box attack means to craft adversarial examples with only the top-1 label of the victim model available. A common practice is to start from a large perturbation and then iteratively reduce it with a deterministic direction and a random one while keeping it adversarial. The limited information obtained from each query and inefficient direction sampling impede attack efficiency, making it hard to obtain a small enough perturbation within a limited number of queries. To tackle this problem, we propose a novel attack method termed Adaptive History-driven Attack (AHA) which gathers information from all historical queries as the prior for current sampling. Moreover, to balance between the deterministic direction and the random one, we dynamically adjust the coefficient according to the ratio of the actual magnitude reduction to the expected one. Such a strategy improves the success rate of queries during optimization, letting adversarial examples move swiftly along the decision boundary. Our method can also integrate with subspace optimization like dimension reduction to further improve efficiency. Extensive experiments on both ImageNet and CelebA datasets demonstrate that our method achieves at least 24.3% lower magnitude of perturbation on average with the same number of queries. Finally, we prove the practical potential of our method by evaluating it on popular defense methods and a real-world system provided by MEGVII Face++.
引用
收藏
页码:16148 / 16157
页数:10
相关论文
共 50 条
  • [11] Zero-Shot Knowledge Distillation from a Decision-Based Black-Box Model
    Wang, Zi
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139 : 7688 - 7699
  • [12] Black-Box Decision based Adversarial Attack with Symmetric α-stable Distribution
    Srinivasan, Vignesh
    Kuruoglu, Ercan E.
    Mueller, Klaus-Robert
    Samek, Wojciech
    Nakajima, Shinichi
    2019 27TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO), 2019,
  • [13] Fooling Decision-Based Black-Box Automotive Vision Perception Systems in Physical World
    Jia, Wei
    Lu, Zhaojun
    Yu, Runze
    Li, Liaoyuan
    Zhang, Haichun
    Liu, Zhenglin
    Qu, Gang
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024, 25 (07) : 7081 - 7092
  • [14] Projection & Probability-Driven Black-Box Attack
    Li, Jie
    Li, Rongrong
    Liu, Hong
    Liu, Jianzhuang
    Zhong, Bineng
    Deng, Cheng
    Tian, Qi
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 359 - 368
  • [15] Adaptive hyperparameter optimization for black-box adversarial attack
    Zhenyu Guan
    Lixin Zhang
    Bohan Huang
    Bihe Zhao
    Song Bian
    International Journal of Information Security, 2023, 22 : 1765 - 1779
  • [16] Adaptive hyperparameter optimization for black-box adversarial attack
    Guan, Zhenyu
    Zhang, Lixin
    Huang, Bohan
    Zhao, Bihe
    Bian, Song
    INTERNATIONAL JOURNAL OF INFORMATION SECURITY, 2023, 22 (06) : 1765 - 1779
  • [17] Black-Box Boundary Attack Based on Gradient Optimization
    Yang, Yuli
    Liu, Zishuo
    Lei, Zhen
    Wu, Shuhong
    Chen, Yongle
    ELECTRONICS, 2024, 13 (06)
  • [18] Reinforcement Learning Based Sparse Black-box Adversarial Attack on Video Recognition Models
    Wang, Zeyuan
    Sha, Chaofeng
    Yang, Su
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 3162 - 3168
  • [19] TrojLLM: A Black-box Trojan Prompt Attack on Large Language Models
    Xue, Jiaqi
    Zheng, Mengxin
    Hua, Ting
    Shen, Yilin
    Liu, Yepeng
    Boloni, Ladislau
    Lou, Qian
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [20] Targeted Black-Box Adversarial Attack Method for Image Classification Models
    Zheng, Su
    Chen, Jialin
    Wang, Lingli
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,