Generating Fluent Chinese Adversarial Examples for Sentiment Classification

被引:0
|
作者
Wang, Congyi [1 ,2 ]
Zeng, Jianping [1 ,2 ]
Wu, Chengrong [1 ,2 ]
机构
[1] Fudan Univ, Sch Comp Sci, Shanghai 200433, Peoples R China
[2] Minist Educ, Engn Res Ctr Cyber Secur Auditing & Monitoring, Shanghai 200433, Peoples R China
基金
国家重点研发计划;
关键词
Adversarial examples; Chinese natural language; Sentiment classification;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Highly accurate classifiers can be trained by existing machine learning models, however, most of these classifiers do not consider the adversarial attack. This makes these classifiers vulnerable to adversarial examples. In order to improve the ability of sentiment classifiers to resist the adversarial attack, it is very important to generate high-quality adversarial examples. Most of the existing methods that generate natural language adversarial examples aim at English text with relatively simple strategies, but a single transformation strategy is easily detected by the defender. In this paper, we propose a new method to generate Chinese natural language adversarial examples, which is called AD-ER (Adversarial Examples with Readability). The first step is to select the important words in the text, which have great impact on the sentiment classifier. Then we proposed four variant strategies to replace the important words and the best candidate word is selected heuristically under the constraints of its readability and maximum entropy model. The simulation results on a real shopping review dataset verify that the examples generated by our method can produce large attack disturbance to the classifiers. Different from other examples, our examples have good readability and diversity, which are more fluent and harder to be detected.
引用
收藏
页码:149 / +
页数:6
相关论文
共 50 条
  • [41] Generating Distributional Adversarial Examples to Evade Statistical Detectors
    Kaya, Yigitcan
    Zafar, Muhammad Bilal
    Aydore, Sergul
    Rauschmayr, Nathalie
    Kenthapadi, Krishnaram
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [42] Generating adversarial examples without specifying a target model
    Yang, Gaoming
    Li, Mingwei
    Fang, Xianjing
    Zhang, Ji
    Liang, Xingzhu
    PEERJ COMPUTER SCIENCE, 2021, 7
  • [43] Generating Robust Audio Adversarial Examples with Temporal Dependency
    Zhang, Hongting
    Zhou, Pan
    Yan, Qiben
    Liu, Xiao-Yang
    PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 3167 - 3173
  • [44] Generating Adversarial Examples through Latent Space Exploration of Generative Adversarial Networks
    Clare, Luana
    Correia, Joao
    PROCEEDINGS OF THE 2023 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE COMPANION, GECCO 2023 COMPANION, 2023, : 1760 - 1767
  • [45] Sentiment classification with adversarial learning and attention mechanism
    Xu, Yueshen
    Li, Lei
    Gao, Honghao
    Hei, Lei
    Li, Rui
    Wang, Yihao
    COMPUTATIONAL INTELLIGENCE, 2021, 37 (02) : 774 - 798
  • [46] Towards robust classification detection for adversarial examples
    Liu, Huangxiaolie
    Zhang, Dong
    Chen, Huijun
    INTERNATIONAL CONFERENCE FOR INTERNET TECHNOLOGY AND SECURED TRANSACTIONS (ICITST-2020), 2020, : 23 - 29
  • [47] Adversarial examples for extreme multilabel text classification
    Mohammadreza Qaraei
    Rohit Babbar
    Machine Learning, 2022, 111 : 4539 - 4563
  • [48] Adversarial examples for extreme multilabel text classification
    Qaraei, Mohammadreza
    Babbar, Rohit
    MACHINE LEARNING, 2022, 111 (12) : 4539 - 4563
  • [49] On Effectiveness of Adversarial Examples and Defenses for Malware Classification
    Podschwadt, Robert
    Takabi, Hassan
    SECURITY AND PRIVACY IN COMMUNICATION NETWORKS, SECURECOMM, PT II, 2019, 305 : 380 - 393
  • [50] Generating Black-Box Adversarial Examples in Sparse Domain
    Zanddizari, Hadi
    Zeinali, Behnam
    Chang, J. Morris
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2022, 6 (04): : 795 - 804