Textual Backdoor Attack via Keyword Positioning

被引:0
|
作者
Chen, Depeng [1 ,2 ]
Mao, Fangfang [1 ]
Jin, Hulin [1 ]
Cui, Jie [1 ,2 ]
机构
[1] Anhui Univ, Sch Comp Sci & Technol, Hefei 230601, Peoples R China
[2] Hefei Comprehens Natl Sci Ctr, Inst Artificial Intelligence, Hefei 230026, Peoples R China
关键词
Backdoor attack; NLP; DNN;
D O I
10.1007/978-981-97-5609-4_5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The backdoor problem poses a potential threat to the security of neural networks. Although backdoor attacks have been extensively studied in the field of CV, they cannot be directly applied to the field of NLP due to the discrete nature of data characteristics. Data poisoning attacks are a common strategy in backdoor attacks in the NLP field, such as replacing or inserting triggers (for example, rare words) into sentences. However, most of the previous work was to randomly select the location of the trigger to be replaced or inserted, and inserting rare words can cause abnormal natural language expression and can be easily detected. In response to the above problems, this paper proposes a textual back door attack technique based on keyword positioning. Keywords usually calculate the importance score of each word or word with a specific part of speech and find the most vulnerable words in the sentence, that is, the keywords that help the target model make judgments. Therefore, interference with these words often makes the target model make errors in judgment. In this article, we first calculate the importance score and part-of-speech label of each word in the sentence, then select the trigger word based on the false correlation between the single word and the target label, and finally perturb the position of the keyword. We conducted experiments on four text classification data sets, and the results showed that the attack we proposed can not only ensure the concealment of the trigger in most cases but also has a better attack than the baseline solution.
引用
收藏
页码:55 / 66
页数:12
相关论文
共 50 条
  • [21] SIMTROJAN: STEALTHY BACKDOOR ATTACK
    Ren, Yankun
    Li, Longfei
    Zhou, Jun
    2021 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2021, : 819 - 823
  • [22] Fake the Real: Backdoor Attack on Deep Speech Classification via Voice Conversion
    Ye, Zhe
    Mao, Terui
    Dong, Li
    Yan, Diqun
    INTERSPEECH 2023, 2023, : 4923 - 4927
  • [23] Hidden Trigger Backdoor Attack on NLP Models via Linguistic Style Manipulation
    Pan, Xudong
    Zhang, Mi
    Sheng, Beina
    Zhu, Jiaming
    Yang, Min
    PROCEEDINGS OF THE 31ST USENIX SECURITY SYMPOSIUM, 2022, : 3611 - 3628
  • [24] Inaudible Backdoor Attack via Stealthy Frequency Trigger Injection in Audio Spectrogram
    Zhang, Tianfang
    Huy Phan
    Tang, Zijie
    Shi, Cong
    Wang, Yan
    Yuan, Bo
    Chen, Yingying
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL CONFERENCE ON MOBILE COMPUTING AND NETWORKING, ACM MOBICOM 2024, 2024, : 31 - 45
  • [25] Mitigating backdoor attacks in LSTM-based text classification systems by Backdoor Keyword Identification
    Chen, Chuanshuai
    Dai, Jiazhu
    NEUROCOMPUTING, 2021, 452 : 253 - 262
  • [26] STEALTHY BACKDOOR ATTACK WITH ADVERSARIAL TRAINING
    Feng, Le
    Li, Sheng
    Qian, Zhenxing
    Zhang, Xinpeng
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 2969 - 2973
  • [27] Backdoor Attack With Sparse and Invisible Trigger
    Gao, Yinghua
    Li, Yiming
    Gong, Xueluan
    Li, Zhifeng
    Xia, Shu-Tao
    Wang, Qian
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 6364 - 6376
  • [28] Spatialspectral-Backdoor: Realizing backdoor attack for deep neural networks in brain-computer interface via EEG characteristics
    Li, Fumin
    Huang, Mengjie
    You, Wenlong
    Zhu, Longsheng
    Cheng, Hanjing
    Yang, Rui
    NEUROCOMPUTING, 2025, 616
  • [29] Sniper Backdoor: Single Client Targeted Backdoor Attack in Federated Learning
    Abad, Gorka
    Paguada, Servio
    Ersoy, Oguzhan
    Picek, Stjepan
    Ramirez-Duran, Victor Julio
    Urbieta, Aitor
    2023 IEEE CONFERENCE ON SECURE AND TRUSTWORTHY MACHINE LEARNING, SATML, 2023, : 377 - 391
  • [30] Data Poisoning Quantization Backdoor Attack
    Tran Huynh
    Anh Tran
    Khoa D Doan
    Tung Pham
    COMPUTER VISION - ECCV 2024, PT LXXXIV, 2025, 15142 : 38 - 54