Improving Adversarial Robustness With Adversarial Augmentations

被引:2
|
作者
Chen, Chuanxi [1 ,2 ]
Ye, Dengpan [1 ,2 ]
He, Yiheng [1 ,2 ]
Tang, Long [1 ,2 ]
Xu, Yue [1 ,2 ]
机构
[1] Wuhan Univ, Key Lab Aerosp Informat Secur & Trusted Comp, Minist Educ, Wuhan 430072, Peoples R China
[2] Wuhan Univ, Sch Cyber Sci & Engn, Wuhan 430072, Peoples R China
基金
中国国家自然科学基金;
关键词
Training; Robustness; Internet of Things; Security; Perturbation methods; Feature extraction; Data augmentation; Adversarial robustness; augmentations; contrastive learning (CL); deep neural networks (DNNs); Internet of Things (IoT) security;
D O I
10.1109/JIOT.2023.3301608
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep neural network (DNN)-based applications are extensively being researched and applied in the Internet of Things (IoT) devices in daily lives due to impressive performance. Recently, adversarial attacks pose a significant threat to the security of deep neural networks (DNNs), adversarial training has emerged as a promising and effective defense approach for defending against such attacks. However, existing adversarial training methods have shown limited success in defending against attacks unseen during training, thereby undermining their effectiveness. Besides, generating adversarial perturbations for adversarial training requires massive expensive labeled data, which is a critical obstacle in the robust DNNs-based IoT applications. In this article, we first explore the effective data augmentations by implementing adversarial attacks with self-supervised in latent space. Then, we propose new loss metric functions that can avoid collapse phenomenon of contrastive learning (CL) by measuring the distances between adversarial augmented pairs. Based on the extracted adversarial features in self-supervised CL, we propose a novel adversarial robust learning (ARL) method, which implements adversarial training without any labels and obtains more general robust encoder network. Our approach is validated on commonly used benchmark data sets and models, where it achieves comparable adversarial robustness against different adversarial attacks when compared to supervised adversarial training methods. Additionally, ARL outperforms state-of-the-art self-supervised adversarial learning techniques in terms of achieving higher robustness and clean prediction accuracy for the downstream classification task.
引用
收藏
页码:5105 / 5117
页数:13
相关论文
共 50 条
  • [41] Improving adversarial robustness of traffic sign image recognition networks
    Hashemi, Atiye Sadat
    Mozaffari, Saeed
    Alirezaee, Shahpour
    DISPLAYS, 2022, 74
  • [42] Improving the Robustness of the Bug Triage Model through Adversarial Training
    Kim, Min-ha
    Wang, Dae-sung
    Wang, Sheng-tsai
    Park, Seo-Hyeon
    Lee, Chan-gun
    36TH INTERNATIONAL CONFERENCE ON INFORMATION NETWORKING (ICOIN 2022), 2022, : 478 - 481
  • [43] Improving the Robustness of Model Compression by On-Manifold Adversarial Training
    Kwon, Junhyung
    Lee, Sangkyun
    FUTURE INTERNET, 2021, 13 (12)
  • [44] Improving Adversarial Robustness via Distillation-Based Purification
    Koo, Inhwa
    Chae, Dong-Kyu
    Lee, Sang-Chul
    Cascio, Donato
    APPLIED SCIENCES-BASEL, 2023, 13 (20):
  • [45] Improving DNN Robustness to Adversarial Attacks Using Jacobian Regularization
    Jakubovitz, Daniel
    Girye, Raja
    COMPUTER VISION - ECCV 2018, PT XII, 2018, 11216 : 525 - 541
  • [46] Robust Proxy: Improving Adversarial Robustness by Robust Proxy Learning
    Lee, Hong Joo
    Ro, Yong Man
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2023, 18 : 4021 - 4033
  • [47] SPLASH: Learnable activation functions for improving accuracy and adversarial robustness
    Tavakoli, Mohammadamin
    Agostinelli, Forest
    Baldi, Pierre
    NEURAL NETWORKS, 2021, 140 : 1 - 12
  • [48] AugMax: Adversarial Composition of Random Augmentations for Robust Training
    Wang, Haotao
    Xiao, Chaowei
    Kossaifi, Jean
    Yu, Zhiding
    Anandkumar, Anima
    Wang, Zhangyang
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [49] Poster: Boosting Adversarial Robustness by Adversarial Pre-training
    Xu, Xiaoyun
    Picek, Stjepan
    PROCEEDINGS OF THE 2023 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, CCS 2023, 2023, : 3540 - 3542
  • [50] Improving the Transferability of Adversarial Samples with Adversarial Transformations
    Wu, Weibin
    Su, Yuxin
    Lyu, Michael R.
    King, Irwin
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 9020 - 9029