A Practical Black-Box Attack on Source Code Authorship Identification Classifiers

被引:7
|
作者
Liu, Qianjun [1 ]
Ji, Shouling [1 ]
Liu, Changchang [2 ]
Wu, Chunming [1 ]
机构
[1] Zhejiang Univ, Coll Comp Sci & Technol, Hangzhou 310027, Peoples R China
[2] IBM Thomas J Watson Res Ctr, Dept Distributed AI, Yorktown Hts, NY 10598 USA
基金
中国国家自然科学基金;
关键词
Feature extraction; Tools; Training; Syntactics; Predictive models; Perturbation methods; Transforms; Source code; authorship identification; adversarial stylometry; ROBUSTNESS;
D O I
10.1109/TIFS.2021.3080507
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Existing researches have recently shown that adversarial stylometry of source code can confuse source code authorship identification (SCAI) models, which may threaten the security of related applications such as programmer attribution, software forensics, etc. In this work, we propose source code authorship disguise (SCAD) to automatically hide programmers' identities from authorship identification, which is more practical than the previous work that requires to known the output probabilities or internal details of the target SCAI model. Specifically, SCAD trains a substitute model and develops a set of semantically equivalent transformations, based on which the original code is modified towards a disguised style with small manipulations in lexical features and syntactic features. When evaluated under totally black-box settings, on a real-world dataset consisting of 1,600 programmers, SCAD induces state-of-the-art SCAI models to cause above 30% misclassification rates. The efficiency and utility-preserving properties of SCAD are also demonstrated with multiple metrics. Furthermore, our work can serve as a guideline for developing more robust identification methods in the future.
引用
下载
收藏
页码:3620 / 3633
页数:14
相关论文
共 50 条
  • [1] Black-box Universal Adversarial Attack on Text Classifiers
    Zhang, Yu
    Shao, Kun
    Yang, Junan
    Liu, Hui
    2021 2ND ASIA CONFERENCE ON COMPUTERS AND COMMUNICATIONS (ACCC 2021), 2021, : 1 - 5
  • [2] Black-Box Adversarial Attack for Deep Learning Classifiers in IoT Applications
    Singh, Abhijit
    Sikdar, Biplab
    2022 IEEE 8TH WORLD FORUM ON INTERNET OF THINGS, WF-IOT, 2022,
  • [3] An Evolutionary-Based Black-Box Attack to Deep Neural Network Classifiers
    Zhou, Yutian
    Tan, Yu-an
    Zhang, Quanxin
    Kuang, Xiaohui
    Han, Yahong
    Hu, Jingjing
    MOBILE NETWORKS & APPLICATIONS, 2021, 26 (04): : 1616 - 1629
  • [4] An Evolutionary-Based Black-Box Attack to Deep Neural Network Classifiers
    Yutian Zhou
    Yu-an Tan
    Quanxin Zhang
    Xiaohui Kuang
    Yahong Han
    Jingjing Hu
    Mobile Networks and Applications, 2021, 26 : 1616 - 1629
  • [5] Data-Free Adversarial Perturbations for Practical Black-Box Attack
    Huan, Zhaoxin
    Wang, Yulong
    Zhang, Xiaolu
    Shang, Lin
    Fu, Chilin
    Zhou, Jun
    ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2020, PT II, 2020, 12085 : 127 - 138
  • [6] A Practical Black-Box Attack Against Autonomous Speech Recognition Model
    Fan, Wenshu
    Li, Hongwei
    Jiang, Wenbo
    Xu, Guowen
    Lu, Rongxing
    2020 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2020,
  • [7] A black-box adversarial attack strategy with adjustable sparsity and generalizability for deep image classifiers
    Ghosh, Arka
    Mullick, Sankha Subhra
    Datta, Shounak
    Das, Swagatam
    Das, Asit Kr
    Mallipeddi, Rammohan
    PATTERN RECOGNITION, 2022, 122
  • [8] Active Bayesian Assessment of Black-Box Classifiers
    Ji, Disi
    Logan, Robert L.
    Smyth, Padhraic
    Steyvers, Mark
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 7935 - 7944
  • [9] Generative causal explanations of black-box classifiers
    O'Shaughnessy, Matthew
    Canal, Gregory
    Connor, Marissa
    Davenport, Mark
    Rozell, Christopher
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [10] Explaining black-box classifiers: Properties and functions
    Amgoud, Leila
    INTERNATIONAL JOURNAL OF APPROXIMATE REASONING, 2023, 155 : 40 - 65