Projan: A probabilistic trojan attack on deep neural networks

被引:0
|
作者
Saremi, Mehrin [1 ]
Khalooei, Mohammad [2 ]
Rastgoo, Razieh [3 ]
Sabokrou, Mohammad [4 ,5 ]
机构
[1] Semnan University, Farzanegan Campus, Semnan,35131-19111, Iran
[2] Amirkabir University of Technology, Department of Computer Engineering, Tehran, Iran
[3] Faculty of Electrical and Computer Engineering, Semnan University, Semnan,35131-19111, Iran
[4] Institute for Research in Fundamental Sciences, Tehran, Iran
[5] Okinawa Institute of Science and Technology, Okinawa, Japan
关键词
D O I
10.1016/j.knosys.2024.112565
中图分类号
学科分类号
摘要
Deep neural networks have gained popularity due to their outstanding performance across various domains. However, because of their lack of explainability, they are vulnerable to some kinds of threats including the trojan or backdoor attack, in which an adversary can train the model to respond to a crafted peculiar input pattern (also called trigger) according to their will. Several trojan attack and defense methods have been proposed in the literature. Many of the defense methods are based on the assumption that the possibly existing trigger must be able to affect the model's behavior, making it output a certain class label for all inputs. In this work, we propose an alternative attack method that violates this assumption. Instead of a single trigger that works on all inputs, a few triggers are generated that will affect only some of the inputs. At attack time, the adversary will need to try more than one trigger to succeed, which might be possible in some real-world situations. Our experiments on MNIST and CIFAR-10 datasets show that such an attack can be implemented successfully, reaching an attack success rate similar to baseline methods called BadNet and N-to-One. We also tested wide range of defense methods and verified that in general, this kind of backdoor is more difficult for defense algorithms to detect. The code is available at https://github.com/programehr/Projan. © 2024 Elsevier B.V.
引用
收藏
相关论文
共 50 条
  • [41] SPA: An Efficient Adversarial Attack on Spiking Neural Networks using Spike Probabilistic
    Lin, Xuanwei
    Dong, Chen
    Liu, Ximeng
    Zhang, Yuanyuan
    2022 22ND IEEE/ACM INTERNATIONAL SYMPOSIUM ON CLUSTER, CLOUD AND INTERNET COMPUTING (CCGRID 2022), 2022, : 366 - 375
  • [42] Design and Evaluation of a Multi-Domain Trojan Detection Method on Deep Neural Networks
    Gao, Yansong
    Kim, Yeonjae
    Doan, Bao Gia
    Zhang, Zhi
    Zhang, Gongxuan
    Nepal, Surya
    Ranasinghe, Damith C.
    Kim, Hyoungshick
    IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, 2022, 19 (04) : 2349 - 2364
  • [43] DeepInspect: A Black-box Trojan Detection and Mitigation Framework for Deep Neural Networks
    Chen, Huili
    Fu, Cheng
    Zhao, Jishen
    Koushanfar, Farinaz
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 4658 - 4664
  • [44] Post-stack seismic inversion through probabilistic neural networks and deep forward neural networks
    Sotelo, Victor
    Almanza, Ovidio
    Montes, Luis
    EARTH SCIENCE INFORMATICS, 2024, 17 (03) : 1957 - 1966
  • [45] Backdoor Attack on Deep Neural Networks Triggered by Fault Injection Attack on Image Sensor Interface
    Oyama, Tatsuya
    Okura, Shunsuke
    Yoshida, Kota
    Fujino, Takeshi
    SENSORS, 2023, 23 (10)
  • [46] SGBA: A stealthy scapegoat backdoor attack against deep neural networks
    He, Ying
    Shen, Zhili
    Xia, Chang
    Hua, Jingyu
    Tong, Wei
    Zhong, Sheng
    COMPUTERS & SECURITY, 2024, 136
  • [47] DeepRover: A Query-Efficient Blackbox Attack for Deep Neural Networks
    Zhang, Fuyuan
    Hu, Xinwen
    Ma, Lei
    Zhao, Jianjun
    PROCEEDINGS OF THE 31ST ACM JOINT MEETING EUROPEAN SOFTWARE ENGINEERING CONFERENCE AND SYMPOSIUM ON THE FOUNDATIONS OF SOFTWARE ENGINEERING, ESEC/FSE 2023, 2023, : 1384 - 1394
  • [48] PROBABILISTIC NEURAL NETWORKS
    SPECHT, DF
    NEURAL NETWORKS, 1990, 3 (01) : 109 - 118
  • [49] Zero-Knowledge Attack for Replicating Protected Deep Neural Networks
    Mosafi, Itay
    David, Eli
    Netanyahu, Nathan S.
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [50] Priority Adversarial Example in Evasion Attack on Multiple Deep Neural Networks
    Kwon, Hyun
    Yoon, Hyunsoo
    Choi, Daeseon
    2019 1ST INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE IN INFORMATION AND COMMUNICATION (ICAIIC 2019), 2019, : 399 - 404