A Survey of Adversarial Defenses and Robustness in NLP

被引:36
|
作者
Goyal, Shreya [1 ]
Doddapaneni, Sumanth [1 ]
Khapra, Mitesh M. [1 ]
Ravindran, Balaraman [1 ]
机构
[1] Indian Inst Technol Madras, Bhupat & Jyoti Mehta Sch Biosci, Robert Bosch Ctr Data Sci & AI, Chennai 600036, Tamil Nadu, India
关键词
Adversarial attacks; adversarial defenses; perturbations; NLP; DEEP NEURAL-NETWORKS; COMPUTER VISION; ATTACKS;
D O I
10.1145/3593042
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
In the past few years, it has become increasingly evident that deep neural networks are not resilient enough to withstand adversarial perturbations in input data, leaving them vulnerable to attack. Various authors have proposed strong adversarial attacks for computer vision and Natural Language Processing (NLP) tasks. As a response, many defense mechanisms have also been proposed to prevent these networks from failing. The significance of defending neural networks against adversarial attacks lies in ensuring that the model's predictions remain unchanged even if the input data is perturbed. Several methods for adversarial defense in NLP have been proposed, catering to different NLP tasks such as text classification, named entity recognition, and natural language inference. Some of these methods not only defend neural networks against adversarial attacks but also act as a regularization mechanism during training, saving the model from overfitting. This survey aims to review the various methods proposed for adversarial defenses in NLP over the past few years by introducing a novel taxonomy. The survey also highlights the fragility of advanced deep neural networks in NLP and the challenges involved in defending them.
引用
收藏
页数:39
相关论文
共 50 条
  • [21] CAT-Gen: Improving Robustness in NLP Models via Controlled Adversarial Text Generation
    Wang, Tianlu
    Wang, Xuezhi
    Qin, Yao
    Ben Packer
    Lee, Kang
    Chen, Jilin
    Beutel, Alex
    Chi, Ed
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 5141 - 5146
  • [22] Adversarial attacks and defenses on text-to-image diffusion models: A survey
    Zhang, Chenyu
    Hu, Mingwang
    Li, Wenhui
    Wang, Lanjun
    INFORMATION FUSION, 2025, 114
  • [23] How Deep Learning Sees the World: A Survey on Adversarial Attacks & Defenses
    Costa, Joana C.
    Roxo, Tiago
    Proenca, Hugo
    Inacio, Pedro Ricardo Morais
    IEEE ACCESS, 2024, 12 : 61113 - 61136
  • [24] A survey on adversarial attacks and defenses for object detection and their applications in autonomous vehicles
    Amirkhani, Abdollah
    Karimi, Mohammad Parsa
    Banitalebi-Dehkordi, Amin
    VISUAL COMPUTER, 2023, 39 (11): : 5293 - 5307
  • [25] A survey on adversarial attacks and defenses for object detection and their applications in autonomous vehicles
    Abdollah Amirkhani
    Mohammad Parsa Karimi
    Amin Banitalebi-Dehkordi
    The Visual Computer, 2023, 39 : 5293 - 5307
  • [26] A Survey on Adversarial Deep Learning Robustness in Medical Image Analysis
    Apostolidis, Kyriakos D.
    Papakostas, George A.
    ELECTRONICS, 2021, 10 (17)
  • [27] It Is All about Data: A Survey on the Effects of Data on Adversarial Robustness
    Xiong, Peiyu
    Tegegn, Michael
    Sarin, Jaskeerat Singh
    Pal, Shubhraneel
    Rubin, Julia
    ACM COMPUTING SURVEYS, 2024, 56 (07)
  • [28] Adversarial Attacks and Defenses on 3D Point Cloud Classification: A Survey
    Naderi, Hanieh
    Bajic, Ivan V.
    IEEE ACCESS, 2023, 11 : 144274 - 144295
  • [29] Scaling provable adversarial defenses
    Wong, Eric
    Schmidt, Frank R.
    Metzen, Jan Hendrik
    Kolter, J. Zico
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [30] Adversarial Robustness of Neural Networks from the Perspective of Lipschitz Calculus: A Survey
    Zuehlke, Monty-maximilian
    Kudenko, Daniel
    ACM COMPUTING SURVEYS, 2025, 57 (06)