Pruning Adversarially Robust Neural Networks without Adversarial Examples

被引:1
|
作者
Jian, Tong [1 ]
Wang, Zifeng [1 ]
Wang, Yanzhi [1 ]
Dy, Jennifer [1 ]
Ioannidis, Stratis [1 ]
机构
[1] Northeastern Univ, Dept Elect & Comp Engn, Boston, MA 02115 USA
基金
美国国家科学基金会;
关键词
Adversarial Robustness; Adversarial Pruning; Self-distillation; HSIC Bottleneck;
D O I
10.1109/ICDM54844.2022.00120
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Adversarial pruning compresses models while preserving robustness. Current methods require access to adversarial examples during pruning. This significantly hampers training efficiency. Moreover, as new adversarial attacks and training methods develop at a rapid rate, adversarial pruning methods need to be modified accordingly to keep up. In this work, we propose a novel framework to prune a previously trained robust neural network while maintaining adversarial robustness, without further generating adversarial examples. We leverage concurrent self-distillation and pruning to preserve knowledge in the original model as well as regularizing the pruned model via the HilbertSchmidt Information Bottleneck. We comprehensively evaluate our proposed framework and show its superior performance in terms of both adversarial robustness and efficiency when pruning architectures trained on the MNIST, CIFAR-10, and CIFAR-100 datasets against five state-of-the-art attacks..
引用
收藏
页码:993 / 998
页数:6
相关论文
共 50 条
  • [41] Pruning of generative adversarial neural networks for medical imaging diagnostics with evolution strategy
    Fernandes Jr, Francisco Erivaldo
    Yen, Gary G.
    [J]. INFORMATION SCIENCES, 2021, 558 : 91 - 102
  • [42] Detecting adversarial examples via prediction difference for deep neural networks
    Guo, Feng
    Zhao, Qingjie
    Li, Xuan
    Kuang, Xiaohui
    Zhang, Jianwei
    Han, Yahong
    Tan, Yu-an
    [J]. INFORMATION SCIENCES, 2019, 501 : 182 - 192
  • [43] Detecting Adversarial Examples in Deep Neural Networks using Normalizing Filters
    Gu, Shuangchi
    Yi, Ping
    Zhu, Ting
    Yao, Yao
    Wang, Wei
    [J]. PROCEEDINGS OF THE 11TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE (ICAART), VOL 2, 2019, : 164 - 173
  • [44] Natural Scene Statistics for Detecting Adversarial Examples in Deep Neural Networks
    Kherchouche, Anouar
    Fezza, Sid Ahmed
    Hamidouche, Wassim
    Deforges, Olivier
    [J]. 2020 IEEE 22ND INTERNATIONAL WORKSHOP ON MULTIMEDIA SIGNAL PROCESSING (MMSP), 2020,
  • [45] Digital Watermark Perturbation for Adversarial Examples to Fool Deep Neural Networks
    Feng, Shiyu
    Feng, Feng
    Xu, Xiao
    Wang, Zheng
    Hu, Yining
    Xie, Lizhe
    [J]. 2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [46] Towards Robust Detection of Adversarial Examples
    Pang, Tianyu
    Du, Chao
    Dong, Yinpeng
    Zhu, Jun
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [47] Towards the Development of Robust Deep Neural Networks in Adversarial Settings
    Huster, Todd P.
    Chiang, Cho-Yu Jason
    Chadha, Ritu
    Swami, Ananthram
    [J]. 2018 IEEE MILITARY COMMUNICATIONS CONFERENCE (MILCOM 2018), 2018, : 419 - 424
  • [48] Exploring misclassifications of robust neural networks to enhance adversarial attacks
    Leo Schwinn
    René Raab
    An Nguyen
    Dario Zanca
    Bjoern Eskofier
    [J]. Applied Intelligence, 2023, 53 : 19843 - 19859
  • [49] Fast Training of Deep Neural Networks Robust to Adversarial Perturbations
    Goodwin, Justin
    Brown, Olivia
    Helus, Victoria
    [J]. 2020 IEEE HIGH PERFORMANCE EXTREME COMPUTING CONFERENCE (HPEC), 2020,
  • [50] Exploring misclassifications of robust neural networks to enhance adversarial attacks
    Schwinn, Leo
    Raab, Rene
    Nguyen, An
    Zanca, Dario
    Eskofier, Bjoern
    [J]. APPLIED INTELLIGENCE, 2023, 53 (17) : 19843 - 19859