A Practical Black-Box Attack on Source Code Authorship Identification Classifiers

被引:7
|
作者
Liu, Qianjun [1 ]
Ji, Shouling [1 ]
Liu, Changchang [2 ]
Wu, Chunming [1 ]
机构
[1] Zhejiang Univ, Coll Comp Sci & Technol, Hangzhou 310027, Peoples R China
[2] IBM Thomas J Watson Res Ctr, Dept Distributed AI, Yorktown Hts, NY 10598 USA
基金
中国国家自然科学基金;
关键词
Feature extraction; Tools; Training; Syntactics; Predictive models; Perturbation methods; Transforms; Source code; authorship identification; adversarial stylometry; ROBUSTNESS;
D O I
10.1109/TIFS.2021.3080507
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Existing researches have recently shown that adversarial stylometry of source code can confuse source code authorship identification (SCAI) models, which may threaten the security of related applications such as programmer attribution, software forensics, etc. In this work, we propose source code authorship disguise (SCAD) to automatically hide programmers' identities from authorship identification, which is more practical than the previous work that requires to known the output probabilities or internal details of the target SCAI model. Specifically, SCAD trains a substitute model and develops a set of semantically equivalent transformations, based on which the original code is modified towards a disguised style with small manipulations in lexical features and syntactic features. When evaluated under totally black-box settings, on a real-world dataset consisting of 1,600 programmers, SCAD induces state-of-the-art SCAI models to cause above 30% misclassification rates. The efficiency and utility-preserving properties of SCAD are also demonstrated with multiple metrics. Furthermore, our work can serve as a guideline for developing more robust identification methods in the future.
引用
下载
收藏
页码:3620 / 3633
页数:14
相关论文
共 50 条
  • [11] SIMULATOR ATTACK plus FOR BLACK-BOX ADVERSARIAL ATTACK
    Ji, Yimu
    Ding, Jianyu
    Chen, Zhiyu
    Wu, Fei
    Zhang, Chi
    Sun, Yiming
    Sun, Jing
    Liu, Shangdong
    2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 636 - 640
  • [12] Perceptual quality-preserving black-box attack against deep learning image classifiers
    Gragnaniello, Diego
    Marra, Francesco
    Verdoliva, Luisa
    Poggi, Giovanni
    PATTERN RECOGNITION LETTERS, 2021, 147 : 142 - 149
  • [13] Query-Efficient Black-Box Attack Against Sequence-Based Malware Classifiers
    Rosenberg, Ishai
    Shabtai, Asaf
    Elovici, Yuval
    Rokach, Lior
    36TH ANNUAL COMPUTER SECURITY APPLICATIONS CONFERENCE (ACSAC 2020), 2020, : 611 - 626
  • [14] ABCAttack: A Gradient-Free Optimization Black-Box Attack for Fooling Deep Image Classifiers
    Cao, Han
    Si, Chengxiang
    Sun, Qindong
    Liu, Yanxiao
    Li, Shancang
    Gope, Prosanta
    ENTROPY, 2022, 24 (03)
  • [15] DIP: Dead code Insertion based Black-box Attack for Programming Language Model
    Na, CheolWon
    Choi, YunSeok
    Lee, Jee-Hyong
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 7777 - 7791
  • [16] Evading Black-box Classifiers Without Breaking Eggs
    Debenedetti, Edoardo
    Carlini, Nicholas
    Tramer, Florian
    IEEE CONFERENCE ON SAFE AND TRUSTWORTHY MACHINE LEARNING, SATML 2024, 2024, : 408 - 424
  • [17] Amora: Black-box Adversarial Morphing Attack
    Wang, Run
    Juefei-Xu, Felix
    Guo, Qing
    Huang, Yihao
    Xie, Xiaofei
    Ma, Lei
    Liu, Yang
    MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, 2020, : 1376 - 1385
  • [18] Fairness improvement for black-box classifiers with Gaussian process
    Dang Nguyen
    Gupta, Sunil
    Rana, Santu
    Shilton, Alistair
    Venkatesh, Svetha
    INFORMATION SCIENCES, 2021, 576 : 542 - 556
  • [19] Automated Image Reduction for Explaining Black-box Classifiers
    Jiang, Mingyue
    Tang, Chengjian
    Zhang, Xiao-Yi
    Zhao, Yangyang
    Ding, Zuohua
    2023 IEEE INTERNATIONAL CONFERENCE ON SOFTWARE ANALYSIS, EVOLUTION AND REENGINEERING, SANER, 2023, : 367 - 378
  • [20] Adversarial Eigen Attack on Black-Box Models
    Zhou, Linjun
    Cui, Peng
    Zhang, Xingxuan
    Jiang, Yinan
    Yang, Shiqiang
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 15233 - 15241