Defending Against Universal Perturbations With Shared Adversarial Training

被引:27
|
作者
Mummadi, Chaithanya Kumar [1 ,2 ]
Brox, Thomas [1 ]
Metzen, Jan Hendrik [2 ]
机构
[1] Univ Freiburg, Freiburg, Germany
[2] Bosch Ctr Artificial Intelligence, Stuttgart, Germany
关键词
D O I
10.1109/ICCV.2019.00503
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Classifiers such as deep neural networks have been shown to be vulnerable against adversarial perturbations on problems with high-dimensional input space. While adversarial training improves the robustness of image classifiers against such adversarial perturbations, it leaves them sensitive to perturbations on a non-negligible fraction of the inputs. In this work, we show that adversarial training is more effective in preventing universal perturbations, where the same perturbation needs to fool a classifier on many inputs. Moreover, we investigate the trade-off between robustness against universal perturbations and performance on unperturbed data and propose an extension of adversarial training that handles this trade-off more gracefully. We present results for image classification and semantic segmentation to showcase that universal perturbations that fool a model hardened with adversarial training become clearly perceptible and show patterns of the target scene.
引用
下载
收藏
页码:4927 / 4936
页数:10
相关论文
共 50 条
  • [21] Defending Person Detection Against Adversarial Patch Attack by Using Universal Defensive Frame
    Yu, Youngjoon
    Lee, Hong Joo
    Lee, Hakmin
    Ro, Yong Man
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2022, 31 : 6976 - 6990
  • [22] Universal adversarial perturbations generative network
    Zheng Wang
    Yang Yang
    Jingjing Li
    Xiaofeng Zhu
    World Wide Web, 2022, 25 : 1725 - 1746
  • [23] Universal adversarial perturbations generative network
    Wang, Zheng
    Yang, Yang
    Li, Jingjing
    Zhu, Xiaofeng
    WORLD WIDE WEB-INTERNET AND WEB INFORMATION SYSTEMS, 2022, 25 (04): : 1725 - 1746
  • [24] Universal Adversarial Training
    Shafahi, Ali
    Najibi, Mahyar
    Xu, Zheng
    Dickerson, John
    Davis, Larry S.
    Goldstein, Tom
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 5636 - 5643
  • [25] A Neural Rejection System Against Universal Adversarial Perturbations in Radio Signal Classification
    Zhang, Lu
    Lambotharan, Sangarapillai
    Zheng, Gan
    Roli, Fabio
    2021 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2021,
  • [26] Guarding Against Universal Adversarial Perturbations in Data-driven Cloud/Edge Services
    Zhou, Xingyu
    Canady, Robert
    Li, Yi
    Bao, Shunxing
    Barve, Yogesh
    Balasubramanian, Daniel
    Gokhale, Aniruddha
    2022 IEEE INTERNATIONAL CONFERENCE ON CLOUD ENGINEERING (IC2E 2022), 2022, : 233 - 244
  • [27] Universal Adversarial Perturbations for Speech Recognition Systems
    Neekhara, Paarth
    Hussain, Shehzeen
    Pandey, Prakhar
    Dubnov, Shlomo
    McAuley, Julian
    Koushanfar, Farinaz
    INTERSPEECH 2019, 2019, : 481 - 485
  • [28] Universal adversarial examples and perturbations for quantum classifiers
    Weiyuan Gong
    Dong-Ling Deng
    National Science Review, 2022, 9 (06) : 48 - 55
  • [29] Universal adversarial examples and perturbations for quantum classifiers
    Gong, Weiyuan
    Deng, Dong-Ling
    NATIONAL SCIENCE REVIEW, 2022, 9 (06)
  • [30] Learning Universal Adversarial Perturbations with Generative Models
    Hayes, Jamie
    Danezis, George
    2018 IEEE SYMPOSIUM ON SECURITY AND PRIVACY WORKSHOPS (SPW 2018), 2018, : 43 - 49