Combining Defences Against Data-Poisoning Based Backdoor Attacks on Neural Networks

被引:1
|
作者
Milakovic, Andrea [1 ]
Mayer, Rudolf [1 ,2 ]
机构
[1] Vienna Univ Technol, Favoritenstr 9-11, Vienna, Austria
[2] SBA Res gGmbH, Floragasse 7, Vienna, Austria
关键词
Machine Learning; Poisoning Attacks; Defences;
D O I
10.1007/978-3-031-10684-2_3
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Machine learning-based systems are increasingly used in critical applications such as medical diagnosis, automotive vehicles, or biometric authentication. Because of their importance, they can become the target of various attacks. In a data poisoning attack, the attacker carefully manipulates some input data, e.g. by superimposing a pattern, e.g. to insert a backdoor (a wrong association of the specific pattern to a desired target) into the model during the training phase. This can later be exploited to control the model behaviour during prediction, and attack its integrity, e.g. by identifying someone as the wrong user or not correctly identifying a traffic sign, thus causing road incidents. Poisoning of the training data is difficult to detect, as often, only small amounts of the data need to be manipulated to achieve a successful attack. The backdoors inserted into the model are hard to detect as well, as its unexpected behaviour manifests only when the specific backdoor trigger, which is only known to the attacker, is presented. Nonetheless, several defence mechanisms were proposed, and in the right setting, they can yield usable results; however, they still show shortcomings and insufficient effectiveness in several cases. In this work, we thus try to answer the extent to which combinations of these defences can improve their individual effectiveness. To this end, we first build successful attacks for two datasets and investigate factors influencing the attack success. Our evaluation shows a substantial impact of the type of neural network models and datasets on the effectiveness of the defence. We also show that the choice of the backdoor trigger has a big impact on the attack and its success. Finally, our evaluation shows that a combination of defences can improve existing defences in several cases.
引用
收藏
页码:28 / 47
页数:20
相关论文
共 50 条
  • [1] An Overview of Backdoor Attacks Against Deep Neural Networks and Possible Defences
    Guo, Wei
    Tondi, Benedetta
    Barni, Mauro
    [J]. IEEE OPEN JOURNAL OF SIGNAL PROCESSING, 2022, 3 : 261 - 287
  • [2] Verifying Neural Networks Against Backdoor Attacks
    Pham, Long H.
    Sun, Jun
    [J]. COMPUTER AIDED VERIFICATION (CAV 2022), PT I, 2022, 13371 : 171 - 192
  • [3] A defense method against backdoor attacks on neural networks
    Kaviani, Sara
    Shamshiri, Samaneh
    Sohn, Insoo
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2023, 213
  • [4] Modelling Data Poisoning Attacks Against Convolutional Neural Networks
    Jonnalagadda, Annapurna
    Mohanty, Debdeep
    Zakee, Ashraf
    Kamalov, Firuz
    [J]. JOURNAL OF INFORMATION & KNOWLEDGE MANAGEMENT, 2024, 23 (02)
  • [5] Latent Space-Based Backdoor Attacks Against Deep Neural Networks
    Kristanto, Adrian
    Wang, Shuo
    Rudolph, Carsten
    [J]. 2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [6] Application of complex systems in neural networks against Backdoor attacks
    Kaviani, Sara
    Sohn, Insoo
    Liu, Huaping
    [J]. 11TH INTERNATIONAL CONFERENCE ON ICT CONVERGENCE: DATA, NETWORK, AND AI IN THE AGE OF UNTACT (ICTC 2020), 2020, : 57 - 59
  • [7] Certified Robustness of Nearest Neighbors against Data Poisoning and Backdoor Attacks
    Jia, Jinyuan
    Liu, Yupei
    Cao, Xiaoyu
    Gong, Neil Zhenqiang
    [J]. THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 9575 - 9583
  • [8] Watermarking Graph Neural Networks based on Backdoor Attacks
    Xu, Jing
    Koffas, Stefanos
    Ersoy, Oguzhan
    Picek, Stjepan
    [J]. 2023 IEEE 8TH EUROPEAN SYMPOSIUM ON SECURITY AND PRIVACY, EUROS&P, 2023, : 1179 - 1197
  • [9] Depth-2 neural networks under a data-poisoning attack
    Karmakar, Sayar
    Mukherjee, Anirbit
    Papamarkou, Theodore
    [J]. NEUROCOMPUTING, 2023, 532 : 56 - 66
  • [10] Targeted Data Poisoning Attacks Against Continual Learning Neural Networks
    Li, Huayu
    Ditzler, Gregory
    [J]. 2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,