Improving Adversarial Robustness via Information Bottleneck Distillation

被引:0
|
作者
Kuang, Huafeng [1 ]
Liu, Hong [2 ]
Wu, YongJian [3 ]
Satoh, Shin'ichi [2 ]
Ji, Rongrong [1 ]
机构
[1] Xiamen Univ, Key Lab Multimedia Trusted Percept & Efficient Co, Minist Educ China, Xiamen 361005, Peoples R China
[2] Natl Inst Informat, Tokyo 1018430, Japan
[3] Tencent, Youtu Lab, Shanghai 200233, Peoples R China
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Previous studies have shown that optimizing the information bottleneck can significantly improve the robustness of deep neural networks. Our study closely examines the information bottleneck principle and proposes an Information Bottleneck Distillation approach. This specially designed, robust distillation technique utilizes prior knowledge obtained from a robust pre-trained model to boost information bottlenecks. Specifically, we propose two distillation strategies that align with the two optimization processes of the information bottleneck. Firstly, we use a robust soft-label distillation method to increase the mutual information between latent features and output prediction. Secondly, we introduce an adaptive feature distillation method that automatically transfers relevant knowledge from the teacher model to the student model, thereby reducing the mutual information between the input and latent features. We conduct extensive experiments to evaluate our approach's robustness against state-of-the-art adversarial attackers such as PGD-attack and AutoAttack. Our experimental results demonstrate the effectiveness of our approach in significantly improving adversarial robustness. Our code is available at https://github.com/SkyKuang/IBD.
引用
收藏
页数:18
相关论文
共 50 条
  • [1] Improving adversarial robustness using knowledge distillation guided by attention information bottleneck
    Gong, Yuxin
    Wang, Shen
    Yu, Tingyue
    Jiang, Xunzhi
    Sun, Fanghui
    [J]. INFORMATION SCIENCES, 2024, 665
  • [2] Improving the Adversarial Robustness of NLP Models by Information Bottleneck
    Zhang, Cenyuan
    Zhou, Xiang
    Wan, Yixin
    Zheng, Xiaoqing
    Chang, Kai-Wei
    Hsieh, Cho-Jui
    [J]. FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), 2022, : 3588 - 3598
  • [3] Improving Adversarial Robustness via Distillation-Based Purification
    Koo, Inhwa
    Chae, Dong-Kyu
    Lee, Sang-Chul
    Cascio, Donato
    [J]. APPLIED SCIENCES-BASEL, 2023, 13 (20):
  • [4] Improving Adversarial Robustness via Mutual Information Estimation
    Zhou, Dawei
    Wang, Nannan
    Gao, Xinbo
    Han, Bo
    Wang, Xiaoyu
    Zhan, Yibing
    Liu, Tongliang
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [5] IB-RAR: Information Bottleneck as Regularizer for Adversarial Robustness
    Xu, Xiaoyun
    Perin, Guilherme
    Picek, Stjepan
    [J]. 2023 53RD ANNUAL IEEE/IFIP INTERNATIONAL CONFERENCE ON DEPENDABLE SYSTEMS AND NETWORKS WORKSHOPS, DSN-W, 2023, : 129 - 135
  • [6] Revisiting Hilbert-Schmidt Information Bottleneck for Adversarial Robustness
    Wang, Zifeng
    Jian, Tong
    Masoomi, Aria
    Ioannidis, Stratis
    Dy, Jennifer
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [7] Improving adversarial robustness by learning shared information
    Yu, Xi
    Smedemark-Margulies, Niklas
    Aeron, Shuchin
    Koike-Akino, Toshiaki
    Moulin, Pierre
    Brand, Matthew
    Parsons, Kieran
    Wang, Ye
    [J]. PATTERN RECOGNITION, 2023, 134
  • [8] Improving Adversarial Robustness via Attention and Adversarial Logit Pairing
    Li, Xingjian
    Goodman, Dou
    Liu, Ji
    Wei, Tao
    Dou, Dejing
    [J]. FRONTIERS IN ARTIFICIAL INTELLIGENCE, 2022, 4
  • [9] Enhancing Adversarial Transferability via Information Bottleneck Constraints
    Qi, Biqing
    Gao, Junqi
    Liu, Jianxing
    Wu, Ligang
    Zhou, Bowen
    [J]. IEEE SIGNAL PROCESSING LETTERS, 2024, 31 : 1414 - 1418
  • [10] Adversarial Information Bottleneck
    Zhai, Penglong
    Zhang, Shihua
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (01) : 221 - 230