Feature Distillation in Deep Attention Network Against Adversarial Examples

被引:7
|
作者
Chen, Xin [1 ,2 ,3 ,4 ]
Weng, Jian [5 ]
Deng, Xiaoling [1 ,2 ,3 ,4 ]
Luo, Weiqi [5 ]
Lan, Yubin [1 ,2 ,3 ,4 ]
Tian, Qi [6 ]
机构
[1] South China Agr Univ, Coll Elect Engn, Coll Artificial Intelligence, Guangzhou 510642, Peoples R China
[2] Natl Ctr Int Collaborat Res Precis Agr Aviat Pest, Guangzhou 510642, Peoples R China
[3] Guangdong Lab Lingnan Modern Agr, Guangzhou 510642, Peoples R China
[4] Guangdong Engn Technol Res Ctr Smart Agr, Guangzhou 510642, Peoples R China
[5] Jinan Univ, Coll Informat Sci & Technol, Guangzhou 510632, Peoples R China
[6] Huawei Cloud & AI, Shenzhen 518100, Peoples R China
基金
中国国家自然科学基金;
关键词
Perturbation methods; Robustness; Image coding; Training; Neural networks; Task analysis; Research and development; Adversarial examples; deep attention mechanism; frequency decomposition; image classification;
D O I
10.1109/TNNLS.2021.3113342
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNs) are easily fooled by adversarial examples. Most existing defense strategies defend against adversarial examples based on full information of whole images. In reality, one possible reason as to why humans are not sensitive to adversarial perturbations is that the human visual mechanism often concentrates on most important regions of images. A deep attention mechanism has been applied in many computer fields and has achieved great success. Attention modules are composed of an attention branch and a trunk branch. The encoder/decoder architecture in the attention branch has potential of compressing adversarial perturbations. In this article, we theoretically prove that attention modules can compress adversarial perturbations by destroying potential linear characteristics of DNNs. Considering the distribution characteristics of adversarial perturbations in different frequency bands, we design and compare three types of attention modules based on frequency decomposition and reorganization to defend against adversarial examples. Moreover, we find that our designed attention modules can obtain high classification accuracies on clean images by locating attention regions more accurately. Experimental results on the CIFAR and ImageNet dataset demonstrate that frequency reorganization in attention modules can not only achieve good robustness to adversarial perturbations, but also obtain comparable, even higher classification, accuracies on clean images. Moreover, our proposed attention modules can be integrated with existing defense strategies as components to further improve adversarial robustness.
引用
收藏
页码:3691 / 3705
页数:15
相关论文
共 50 条
  • [31] Moving Target Defense for Embedded Deep Visual Sensing against Adversarial Examples
    Song, Qun
    Yan, Zhenyu
    Tan, Rui
    PROCEEDINGS OF THE 17TH CONFERENCE ON EMBEDDED NETWORKED SENSOR SYSTEMS (SENSYS '19), 2019, : 124 - 137
  • [32] DeepMTD: Moving Target Defense for Deep Visual Sensing against Adversarial Examples
    Song, Qun
    Yan, Zhenyu
    Tan, Rui
    ACM TRANSACTIONS ON SENSOR NETWORKS, 2022, 18 (01)
  • [33] Is Deep Learning Safe for Robot Vision? Adversarial Examples against the iCub Humanoid
    Melis, Marco
    Demontis, Ambra
    Biggio, Battista
    Brown, Gavin
    Fumera, Giorgio
    Roli, Fabio
    2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW 2017), 2017, : 751 - 759
  • [34] Generating Adversarial Examples Against Remote Sensing Scene Classification via Feature Approximation
    Zhu, Rui
    Ma, Shiping
    Lian, Jiawei
    He, Linyuan
    Mei, Shaohui
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2024, 17 : 10174 - 10187
  • [35] Complete Defense Framework to Protect Deep Neural Networks against Adversarial Examples
    Sun, Guangling
    Su, Yuying
    Qin, Chuan
    Xu, Wenbo
    Lu, Xiaofeng
    Ceglowski, Andrzej
    MATHEMATICAL PROBLEMS IN ENGINEERING, 2020, 2020
  • [36] DeepMTD: Moving Target Defense for Deep Visual Sensing against Adversarial Examples
    Song, Qun
    Yan, Zhenyu
    Tan, Rui
    ACM Transactions on Sensor Networks, 2021, 18 (01)
  • [37] Natural Black-Box Adversarial Examples against Deep Reinforcement Learning
    Yu, Mengran
    Sun, Shiliang
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 8936 - 8944
  • [38] Adversarial Feature Distillation for Facial Expression Recognition
    Bai, Mengchao
    Jia, Xi
    Xie, Weicheng
    Shen, Linlin
    PRICAI 2019: TRENDS IN ARTIFICIAL INTELLIGENCE, PT III, 2019, 11672 : 80 - 92
  • [39] An Adversarial Feature Distillation Method for Audio Classification
    Gao, Liang
    Mi, Haibo
    Zhu, Boqing
    Feng, Dawei
    Li, Yicong
    Peng, Yuxing
    IEEE ACCESS, 2019, 7 : 105319 - 105330
  • [40] FEATURE ADVERSARIAL DISTILLATION FOR POINT CLOUD CLASSIFICATION
    Lee, YuXing
    Wu, Wei
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 970 - 974