Common knowledge learning for generating transferable adversarial examples

被引:0
|
作者
Yang, Ruijie [1 ]
Guo, Yuanfang [1 ]
Wang, Junfu [1 ]
Zhou, Jiantao [2 ,3 ]
Wang, Yunhong [1 ]
机构
[1] Beihang Univ, Sch Comp Sci & Engn, Lab Intelligent Recognit & Image Proc, Beijing 100191, Peoples R China
[2] Univ Macau, State Key Lab Internet Things Smart City, Macau 999078, Peoples R China
[3] Univ Macau, Dept Comp & Sci, Macau 999078, Peoples R China
基金
中国国家自然科学基金;
关键词
black-box attack; adversarial transferability; deep neural networks;
D O I
10.1007/s11704-024-40533-4
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper focuses on an important type of black-box attacks, i.e., transfer-based adversarial attacks, where the adversary generates adversarial examples using a substitute (source) model and utilizes them to attack an unseen target model, without knowing its information. Existing methods tend to give unsatisfactory adversarial transferability when the source and target models are from different types of DNN architectures (e.g., ResNet-18 and Swin Transformer). In this paper, we observe that the above phenomenon is induced by the output inconsistency problem. To alleviate this problem while effectively utilizing the existing DNN models, we propose a common knowledge learning (CKL) framework to learn better network weights to generate adversarial examples with better transferability, under fixed network architectures. Specifically, to reduce the model-specific features and obtain better output distributions, we construct a multi-teacher framework, where the knowledge is distilled from different teacher architectures into one student network. By considering that the gradient of input is usually utilized to generate adversarial examples, we impose constraints on the gradients between the student and teacher models, to further alleviate the output inconsistency problem and enhance the adversarial transferability. Extensive experiments demonstrate that our proposed work can significantly improve the adversarial transferability.
引用
收藏
页数:14
相关论文
共 50 条
  • [41] Generating adversarial examples with input significance indicator
    Qiu, Xiaofeng
    Zhou, Shuya
    NEUROCOMPUTING, 2020, 394 : 1 - 12
  • [42] Generating Fluent Adversarial Examples for Natural Languages
    Zhang, Huangzhao
    Zhou, Hao
    Miao, Ning
    Li, Lei
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 5564 - 5569
  • [43] Defending against and generating adversarial examples together with generative adversarial networks
    Ying Wang
    Xiao Liao
    Wei Cui
    Yang Yang
    Scientific Reports, 15 (1)
  • [44] Generating Adversarial Examples for Static PE Malware Detector Based on Deep Reinforcement Learning
    Chen, Jun
    Jiang, Jingfei
    Li, Rongchun
    Dou, Yong
    5TH ANNUAL INTERNATIONAL CONFERENCE ON INFORMATION SYSTEM AND ARTIFICIAL INTELLIGENCE (ISAI2020), 2020, 1575
  • [45] Crafting transferable adversarial examples via contaminating the salient feature variance
    Ren, Yuchen
    Zhu, Hegui
    Sui, Xiaoyan
    Liu, Chong
    INFORMATION SCIENCES, 2023, 644
  • [46] Rethinking the optimization objective for transferable adversarial examples from a fuzzy perspective
    Yang, Xiangyuan
    Lin, Jie
    Zhang, Hanlin
    Zhao, Peng
    NEURAL NETWORKS, 2025, 184
  • [47] Evading Defenses to Transferable Adversarial Examples by Translation-Invariant Attacks
    Dong, Yinpeng
    Pang, Tianyu
    Su, Hang
    Zhu, Jun
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 4307 - 4316
  • [48] UCG: A Universal Cross-Domain Generator for Transferable Adversarial Examples
    Li, Zhankai
    Wang, Weiping
    Li, Jie
    Chen, Kai
    Zhang, Shigeng
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 3023 - 3037
  • [49] GNP ATTACK: TRANSFERABLE ADVERSARIAL EXAMPLES VIA GRADIENT NORM PENALTY
    Wu, Tao
    Luo, Tie
    Wunsch, Donald C.
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 3110 - 3114
  • [50] Timing Attack on Random Forests for Generating Adversarial Examples
    Dan, Yuichiro
    Shibahara, Toshiki
    Takahashi, Junko
    ADVANCES IN INFORMATION AND COMPUTER SECURITY (IWSEC 2020), 2020, 12231 : 285 - 302