Feature Distillation in Deep Attention Network Against Adversarial Examples

被引:7
|
作者
Chen, Xin [1 ,2 ,3 ,4 ]
Weng, Jian [5 ]
Deng, Xiaoling [1 ,2 ,3 ,4 ]
Luo, Weiqi [5 ]
Lan, Yubin [1 ,2 ,3 ,4 ]
Tian, Qi [6 ]
机构
[1] South China Agr Univ, Coll Elect Engn, Coll Artificial Intelligence, Guangzhou 510642, Peoples R China
[2] Natl Ctr Int Collaborat Res Precis Agr Aviat Pest, Guangzhou 510642, Peoples R China
[3] Guangdong Lab Lingnan Modern Agr, Guangzhou 510642, Peoples R China
[4] Guangdong Engn Technol Res Ctr Smart Agr, Guangzhou 510642, Peoples R China
[5] Jinan Univ, Coll Informat Sci & Technol, Guangzhou 510632, Peoples R China
[6] Huawei Cloud & AI, Shenzhen 518100, Peoples R China
基金
中国国家自然科学基金;
关键词
Perturbation methods; Robustness; Image coding; Training; Neural networks; Task analysis; Research and development; Adversarial examples; deep attention mechanism; frequency decomposition; image classification;
D O I
10.1109/TNNLS.2021.3113342
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNs) are easily fooled by adversarial examples. Most existing defense strategies defend against adversarial examples based on full information of whole images. In reality, one possible reason as to why humans are not sensitive to adversarial perturbations is that the human visual mechanism often concentrates on most important regions of images. A deep attention mechanism has been applied in many computer fields and has achieved great success. Attention modules are composed of an attention branch and a trunk branch. The encoder/decoder architecture in the attention branch has potential of compressing adversarial perturbations. In this article, we theoretically prove that attention modules can compress adversarial perturbations by destroying potential linear characteristics of DNNs. Considering the distribution characteristics of adversarial perturbations in different frequency bands, we design and compare three types of attention modules based on frequency decomposition and reorganization to defend against adversarial examples. Moreover, we find that our designed attention modules can obtain high classification accuracies on clean images by locating attention regions more accurately. Experimental results on the CIFAR and ImageNet dataset demonstrate that frequency reorganization in attention modules can not only achieve good robustness to adversarial perturbations, but also obtain comparable, even higher classification, accuracies on clean images. Moreover, our proposed attention modules can be integrated with existing defense strategies as components to further improve adversarial robustness.
引用
收藏
页码:3691 / 3705
页数:15
相关论文
共 50 条
  • [1] Feature Distillation: DNN-Oriented JPEG Compression Against Adversarial Examples
    Liu, Zihao
    Liu, Qi
    Liu, Tao
    Xu, Nuo
    Lin, Xue
    Wang, Yanzhi
    Wen, Wujie
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 860 - 868
  • [2] Feature decoupling and interaction network for defending against adversarial examples
    Wang, Weidong
    Li, Zhi
    Liu, Shuaiwei
    Zhang, Li
    Yang, Jin
    Wang, Yi
    IMAGE AND VISION COMPUTING, 2024, 144
  • [3] Adversarial Examples Against Deep Neural Network based Steganalysis
    Zhang, Yiwei
    Zhang, Weiming
    Chen, Kejiang
    Liu, Jiayang
    Liu, Yujia
    Yu, Nenghai
    PROCEEDINGS OF THE 6TH ACM WORKSHOP ON INFORMATION HIDING AND MULTIMEDIA SECURITY (IH&MMSEC'18), 2018, : 67 - 72
  • [4] FePN: A robust feature purification network to defend against adversarial examples
    Cao, Dongliang
    Wei, Kaimin
    Wu, Yongdong
    Zhang, Jilian
    Feng, Bingwen
    Chen, Jinpeng
    COMPUTERS & SECURITY, 2023, 134
  • [5] Adversarial Examples Against the Deep Learning Based Network Intrusion Detection Systems
    Yang, Kaichen
    Liu, Jianqing
    Zhang, Chi
    Fang, Yuguang
    2018 IEEE MILITARY COMMUNICATIONS CONFERENCE (MILCOM 2018), 2018, : 559 - 564
  • [6] Deep neural rejection against adversarial examples
    Angelo Sotgiu
    Ambra Demontis
    Marco Melis
    Battista Biggio
    Giorgio Fumera
    Xiaoyi Feng
    Fabio Roli
    EURASIP Journal on Information Security, 2020
  • [7] Deep neural rejection against adversarial examples
    Sotgiu, Angelo
    Demontis, Ambra
    Melis, Marco
    Biggio, Battista
    Fumera, Giorgio
    Feng, Xiaoyi
    Roli, Fabio
    EURASIP JOURNAL ON INFORMATION SECURITY, 2020, 2020 (01)
  • [8] EFFECTIVENESS OF RANDOM DEEP FEATURE SELECTION FOR SECURING IMAGE MANIPULATION DETECTORS AGAINST ADVERSARIAL EXAMPLES
    Barni, M.
    Nowroozi, E.
    Tondi, B.
    Zhang, B.
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 2977 - 2981
  • [9] Targeted Adversarial Examples Against RF Deep Classifiers
    Kokalj-Filipovic, Silvija
    Miller, Rob
    Morman, Joshua
    PROCEEDINGS OF THE 2019 ACM WORKSHOP ON WIRELESS SECURITY AND MACHINE LEARNING (WISEML '19), 2019, : 6 - 11
  • [10] Feature Squeezing: Detecting Adversarial Examples in Deep Neural Networks
    Xu, Weilin
    Evans, David
    Qi, Yanjun
    25TH ANNUAL NETWORK AND DISTRIBUTED SYSTEM SECURITY SYMPOSIUM (NDSS 2018), 2018,