Interpretability Derived Backdoor Attacks Detection in Deep Neural Networks: Work-in-Progress

被引:0
|
作者
Wen, Xiangyu [1 ]
Jiang, Wei [1 ]
Zhan, Jinyu [1 ]
Wang, Xupeng [1 ]
He, Zhiyuan [1 ]
机构
[1] Univ Elect Sci & Technol China, Sch Informat & Software Engn, Chengdu, Peoples R China
关键词
D O I
10.1109/emsoft51651.2020.9244019
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Backdoor attacks to deep neural networks (DNNs) have received increasing attentions, particularly in applications from edge computing. The detection of backdoor attacks is a challenging task, due to the lack of transparency in DNN. In this paper, we design a novel method to detect backdoor attacks in deep neural networks, which is derived from the interpretability of a DNN. A comprehensive analysis of the critical path in DNN is conducted, based on which two indicators are proposed, including the correlation coefficient and the discrete degree. Conseqently, an efficient backdoor detection algorithm is proposed, which only needs a few runtime images to identify the backdoor attacks. Initial experiments indicated the efficiency.
引用
收藏
页码:13 / 14
页数:2
相关论文
共 50 条
  • [1] Interpretability-Guided Defense Against Backdoor Attacks to Deep Neural Networks
    Jiang, Wei
    Wen, Xiangyu
    Zhan, Jinyu
    Wang, Xupeng
    Song, Ziwei
    [J]. IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2022, 41 (08) : 2611 - 2624
  • [2] Backdoor smoothing: Demystifying backdoor attacks on deep neural networks
    Grosse, Kathrin
    Lee, Taesung
    Biggio, Battista
    Park, Youngja
    Backes, Michael
    Molloy, Ian
    [J]. Computers and Security, 2022, 120
  • [3] Backdoor smoothing: Demystifying backdoor attacks on deep neural networks
    Grosse, Kathrin
    Lee, Taesung
    Biggio, Battista
    Park, Youngja
    Backes, Michael
    Molloy, Ian
    [J]. COMPUTERS & SECURITY, 2022, 120
  • [4] Latent Backdoor Attacks on Deep Neural Networks
    Yao, Yuanshun
    Li, Huiying
    Zheng, Haitao
    Zhao, Ben Y.
    [J]. PROCEEDINGS OF THE 2019 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY (CCS'19), 2019, : 2041 - 2055
  • [5] An Insight into Fault Propagation in Deep Neural Networks: Work-in-Progress
    Sun, Ruoxu
    Zhan, Jinyu
    Jiang, Wei
    [J]. PROCEEDINGS OF THE 2020 INTERNATIONAL CONFERENCE ON EMBEDDED SOFTWARE (EMSOFT), 2020, : 20 - 21
  • [6] Attacking Neural Networks with Neural Networks: Towards Deep Synchronization for Backdoor Attacks
    Guan, Zihan
    Sun, Lichao
    Du, Mengnan
    Liu, Ninghao
    [J]. PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023, 2023, : 608 - 618
  • [7] A Backdoor Embedding Method for Backdoor Detection in Deep Neural Networks
    Liu, Meirong
    Zheng, Hong
    Liu, Qin
    Xing, Xiaofei
    Dai, Yinglong
    [J]. UBIQUITOUS SECURITY, 2022, 1557 : 1 - 12
  • [8] Detection of backdoor attacks using targeted universal adversarial perturbations for deep neural networks
    Qu, Yubin
    Huang, Song
    Chen, Xiang
    Wang, Xingya
    Yao, Yongming
    [J]. JOURNAL OF SYSTEMS AND SOFTWARE, 2024, 207
  • [9] Backdoor Attacks on Image Classification Models in Deep Neural Networks
    Zhang, Quanxin
    Ma, Wencong
    Wang, Yajie
    Zhang, Yaoyuan
    Shi, Zhiwei
    Li, Yuanzhang
    [J]. CHINESE JOURNAL OF ELECTRONICS, 2022, 31 (02) : 199 - 212
  • [10] INVISIBLE AND EFFICIENT BACKDOOR ATTACKS FOR COMPRESSED DEEP NEURAL NETWORKS
    Phan, Huy
    Xie, Yi
    Liu, Jian
    Chen, Yingying
    Yuan, Bo
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 96 - 100