Data poisoning attacks against machine learning algorithms

被引:22
|
作者
Yerlikaya, Fahri Anil [1 ]
Bahtiyar, Serif [1 ]
机构
[1] Istanbul Tech Univ, Dept Comp Engn, TR-34469 Istanbul, Turkey
关键词
Cybersecurity; Machine learning; Adversarial attack; Data poisoning; Label flipping attack;
D O I
10.1016/j.eswa.2022.118101
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
For the past decade, machine learning technology has increasingly become popular and it has been contributing to many areas that have the potential to influence the society considerably. Generally, machine learning is used by various industries to enhance their performances. Moreover, machine learning algorithms are used to solve some hard problems of systems that may contain very critical information. This makes machine learning algorithms a target of adversaries, which is an important problem for systems that use such algorithms. Therefore, it is significant to determine the performance and the robustness of a machine learning algorithm against attacks. In this paper, we analyze empirically the robustness and performances of six machine learning algorithms against two types of adversarial attacks by using four different datasets and three metrics. In our experiments, we analyze the robustness of Support Vector Machine, Stochastic Gradient Descent, Logistic Regression, Random Forest, Gaussian Naive Bayes, and K-Nearest Neighbor algorithms to create learning models. We observe their performances in spam, botnet, malware, and cancer detection datasets when we launch adversarial attacks against these environments. We use data poisoning for manipulating training data during adversarial attacks, which are random label flipping and distance-based label flipping attacks. We analyze the performance of each algorithm for a specific dataset by modifying the amount of poisoned data and analyzing behaviors of accuracy rate, f1-score, and AUC score. Analyses results show that machine learning algorithms have various performance results and robustness under different adversarial attacks. Moreover, machine learning algorithms are affected differently in each stage of an adversarial attacks. Furthermore, the behavior of a machine learning algorithm highly depends on the type of the dataset. On the other hand, some machine learning algorithms have better robustness and performance results against adversarial attacks for almost all datasets.
引用
收藏
页数:20
相关论文
共 50 条
  • [1] Jangseung: A Guardian for Machine Learning Algorithms to Protect Against Poisoning Attacks
    Wolf, Shaya
    Gamboa, Woodrow
    Borowczak, Mike
    [J]. 2021 IEEE INTERNATIONAL SMART CITIES CONFERENCE (ISC2), 2021,
  • [2] Poisoning Attacks Against Machine Learning: Can Machine Learning Be Trustworthy?
    Oprea, Alina
    Singhal, Anoop
    Vassilev, Apostol
    [J]. COMPUTER, 2022, 55 (11) : 94 - 99
  • [3] Data Poisoning Attacks on Federated Machine Learning
    Sun, Gan
    Cong, Yang
    Dong, Jiahua
    Wang, Qiang
    Lyu, Lingjuan
    Liu, Ji
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (13) : 11365 - 11375
  • [4] A Defense Method against Poisoning Attacks on IoT Machine Learning Using Poisonous Data
    Chiba, Tomoki
    Sei, Yuichi
    Tahara, Yasuyuki
    Ohsuga, Akihiko
    [J]. 2020 IEEE THIRD INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND KNOWLEDGE ENGINEERING (AIKE 2020), 2020, : 100 - 107
  • [5] A Countermeasure Method Using Poisonous Data Against Poisoning Attacks on IoT Machine Learning
    Chiba, Tomoki
    Sei, Yuichi
    Tahara, Yasuyuki
    Ohsuga, Akihiko
    [J]. INTERNATIONAL JOURNAL OF SEMANTIC COMPUTING, 2021, 15 (02) : 215 - 240
  • [6] Data Poisoning Attacks Against Federated Learning Systems
    Tolpegin, Vale
    Truex, Stacey
    Gursoy, Mehmet Emre
    Liu, Ling
    [J]. COMPUTER SECURITY - ESORICS 2020, PT I, 2020, 12308 : 480 - 501
  • [7] Model poisoning attacks against distributed machine learning systems
    Tomsett, Richard
    Chan, Kevin
    Chakraborty, Supriyo
    [J]. ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING FOR MULTI-DOMAIN OPERATIONS APPLICATIONS, 2019, 11006
  • [8] Robustness Evaluations of Sustainable Machine Learning Models against Data Poisoning Attacks in the Internet of Things
    Dunn, Corey
    Moustafa, Nour
    Turnbull, Benjamin
    [J]. SUSTAINABILITY, 2020, 12 (16)
  • [9] Poisoning and Evasion Attacks Against Deep Learning Algorithms in Autonomous Vehicles
    Jiang, Wenbo
    Li, Hongwei
    Liu, Sen
    Luo, Xizhao
    Lu, Rongxing
    [J]. IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2020, 69 (04) : 4439 - 4449
  • [10] Adversarial data poisoning attacks against the PC learning algorithm
    Alsuwat, Emad
    Alsuwat, Hatim
    Valtorta, Marco
    Farkas, Csilla
    [J]. INTERNATIONAL JOURNAL OF GENERAL SYSTEMS, 2020, 49 (01) : 3 - 31