Improving adversarial robustness of deep neural networks by using semantic information

被引:11
|
作者
Wang, Lina [1 ]
Chen, Xingshu [1 ,2 ]
Tang, Rui [1 ]
Yue, Yawei [1 ]
Zhu, Yi [2 ]
Zeng, Xuemei [2 ]
Wang, Wei [2 ]
机构
[1] Sichuan Univ, Sch Cyber Sci & Engn, Chengdu 610065, Peoples R China
[2] Sichuan Univ, Cyber Sci Res Inst, Chengdu 610065, Peoples R China
基金
中国国家自然科学基金;
关键词
Adversarial robustness; Semantic information; Region adversarial training; Targeted universal perturbations; CLASSIFICATION;
D O I
10.1016/j.knosys.2021.107141
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The vulnerability of deep neural networks (DNNs) to adversarial attack, which is an attack that can mislead state-of-the-art classifiers into making an incorrect classification with high confidence by deliberately perturbing the original inputs, raises concerns about the robustness of DNNs to such attacks. Adversarial training, which is the main heuristic method for improving adversarial robustness and the first line of defense against adversarial attacks, requires many sample-by-sample calculations to increase training size and is usually insufficiently strong for an entire network. This paper provides a new perspective on the issue of adversarial robustness, one that shifts the focus from the network as a whole to the critical part of the region close to the decision boundary corresponding to a given class. From this perspective, we propose a method to generate a single but image-agnostic adversarial perturbation that carries the semantic information implying the directions to the fragile parts on the decision boundary and causes inputs to be misclassified as a specified target. We call the adversarial training based on such perturbations "region adversarial training" (RAT), which resembles classical adversarial training but is distinguished in that it reinforces the semantic information missing in the relevant regions. Experimental results on the MNIST and CIFAR-10 datasets show that this approach greatly improves adversarial robustness even when a very small dataset from the training data is used; moreover, it can defend against fast gradient sign method, universal perturbation, projected gradient descent, and Carlini and Wagner adversarial attacks, which have a completely different pattern from those encountered by the model during retraining. (C) 2021 Elsevier B.V. All rights reserved.
引用
下载
收藏
页数:14
相关论文
共 50 条
  • [21] Improving the Robustness of Neural Networks Using K-Support Norm Based Adversarial Training
    Akhtar, Sheikh Waqas
    Rehman, Saad
    Akhtar, Mahmood
    Khan, Muazzam A.
    Riaz, Farhan
    Chaudry, Qaiser
    Young, Rupert
    IEEE ACCESS, 2016, 4 : 9501 - 9511
  • [22] Robustness of Sparsely Distributed Representations to Adversarial Attacks in Deep Neural Networks
    Sardar, Nida
    Khan, Sundas
    Hintze, Arend
    Mehra, Priyanka
    ENTROPY, 2023, 25 (06)
  • [23] Enhancing the Robustness of Deep Neural Networks by Meta-Adversarial Training
    Chang, You-Kang
    Zhao, Hong
    Wang, Wei-Jie
    International Journal of Network Security, 2023, 25 (01) : 122 - 130
  • [24] CSTAR: Towards Compact and Structured Deep Neural Networks with Adversarial Robustness
    Phan, Huy
    Yin, Miao
    Sui, Yang
    Yuan, Bo
    Zonouz, Saman
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 2, 2023, : 2065 - 2073
  • [25] MRobust: A Method for Robustness against Adversarial Attacks on Deep Neural Networks
    Liu, Yi-Ling
    Lomuscio, Alessio
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [26] Exploring the Impact of Conceptual Bottlenecks on Adversarial Robustness of Deep Neural Networks
    Rasheed, Bader
    Abdelhamid, Mohamed
    Khan, Adil
    Menezes, Igor
    Khatak, Asad Masood
    IEEE ACCESS, 2024, 12 : 131323 - 131335
  • [27] Improving adversarial robustness of Bayesian neural networks via multi-task adversarial training
    Chen, Xu
    Liu, Chuancai
    Zhao, Yue
    Jia, Zhiyang
    Jin, Ge
    INFORMATION SCIENCES, 2022, 592 : 156 - 173
  • [28] Improving Face Liveness Detection Robustness with Deep Convolutional Generative Adversarial Networks
    Padnevych, Ruslan
    Semedo, David
    Carmo, David
    Magalhaes, Joao
    2022 30TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO 2022), 2022, : 1866 - 1870
  • [29] Improving the Robustness of Deep Neural Networks via Stability Training
    Zheng, Stephan
    Song, Yang
    Leung, Thomas
    Goodfellow, Ian
    2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 4480 - 4488
  • [30] Improving the adversarial robustness of quantized neural networks via exploiting the feature diversity
    Chu, Tianshu
    Fang, Kun
    Yang, Jie
    Huang, Xiaolin
    PATTERN RECOGNITION LETTERS, 2023, 176 : 117 - 122