Improving adversarial robustness of deep neural networks by using semantic information

被引:11
|
作者
Wang, Lina [1 ]
Chen, Xingshu [1 ,2 ]
Tang, Rui [1 ]
Yue, Yawei [1 ]
Zhu, Yi [2 ]
Zeng, Xuemei [2 ]
Wang, Wei [2 ]
机构
[1] Sichuan Univ, Sch Cyber Sci & Engn, Chengdu 610065, Peoples R China
[2] Sichuan Univ, Cyber Sci Res Inst, Chengdu 610065, Peoples R China
基金
中国国家自然科学基金;
关键词
Adversarial robustness; Semantic information; Region adversarial training; Targeted universal perturbations; CLASSIFICATION;
D O I
10.1016/j.knosys.2021.107141
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The vulnerability of deep neural networks (DNNs) to adversarial attack, which is an attack that can mislead state-of-the-art classifiers into making an incorrect classification with high confidence by deliberately perturbing the original inputs, raises concerns about the robustness of DNNs to such attacks. Adversarial training, which is the main heuristic method for improving adversarial robustness and the first line of defense against adversarial attacks, requires many sample-by-sample calculations to increase training size and is usually insufficiently strong for an entire network. This paper provides a new perspective on the issue of adversarial robustness, one that shifts the focus from the network as a whole to the critical part of the region close to the decision boundary corresponding to a given class. From this perspective, we propose a method to generate a single but image-agnostic adversarial perturbation that carries the semantic information implying the directions to the fragile parts on the decision boundary and causes inputs to be misclassified as a specified target. We call the adversarial training based on such perturbations "region adversarial training" (RAT), which resembles classical adversarial training but is distinguished in that it reinforces the semantic information missing in the relevant regions. Experimental results on the MNIST and CIFAR-10 datasets show that this approach greatly improves adversarial robustness even when a very small dataset from the training data is used; moreover, it can defend against fast gradient sign method, universal perturbation, projected gradient descent, and Carlini and Wagner adversarial attacks, which have a completely different pattern from those encountered by the model during retraining. (C) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Towards Improving Robustness of Deep Neural Networks to Adversarial Perturbations
    Amini, Sajjad
    Ghaemmaghami, Shahrokh
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2020, 22 (07) : 1889 - 1903
  • [2] Improving Adversarial Robustness of Deep Neural Networks via Linear Programming
    Tang, Xiaochao
    Yang, Zhengfeng
    Fu, Xuanming
    Wang, Jianlin
    Zeng, Zhenbing
    [J]. THEORETICAL ASPECTS OF SOFTWARE ENGINEERING, TASE 2022, 2022, 13299 : 326 - 343
  • [3] Interpreting and Improving Adversarial Robustness of Deep Neural Networks With Neuron Sensitivity
    Zhang, Chongzhi
    Liu, Aishan
    Liu, Xianglong
    Xu, Yitao
    Yu, Hang
    Ma, Yuqing
    Li, Tianlin
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 1291 - 1304
  • [4] Improving Interpretability of Deep Neural Networks with Semantic Information
    Dong, Yinpeng
    Su, Hang
    Zhu, Jun
    Zhang, Bo
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 975 - 983
  • [5] Improving adversarial robustness of deep neural networks via adaptive margin evolution
    Ma, Linhai
    Liang, Liang
    [J]. NEUROCOMPUTING, 2023, 551
  • [6] Improving the Robustness of Deep Neural Networks via Adversarial Training with Triplet Loss
    Li, Pengcheng
    Yi, Jinfeng
    Zhou, Bowen
    Zhang, Lijun
    [J]. PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 2909 - 2915
  • [7] Improving the Adversarial Robustness and Interpretability of Deep Neural Networks by Regularizing Their Input Gradients
    Ros, Andrew Slavin
    Doshi-Velez, Finale
    [J]. THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 1660 - 1669
  • [8] Adversarial robustness improvement for deep neural networks
    Charis Eleftheriadis
    Andreas Symeonidis
    Panagiotis Katsaros
    [J]. Machine Vision and Applications, 2024, 35
  • [9] Adversarial robustness improvement for deep neural networks
    Eleftheriadis, Charis
    Symeonidis, Andreas
    Katsaros, Panagiotis
    [J]. MACHINE VISION AND APPLICATIONS, 2024, 35 (03)
  • [10] Robustness of deep neural networks in adversarial examples
    [J]. Song, Xiao (songxiao@buaa.edu.cn), 1600, University of Cincinnati (24):