LineVul: A Transformer-based Line-Level Vulnerability Prediction

被引:135
|
作者
Fu, Michael [1 ]
Tantithamthavorn, Chakkrit [1 ]
机构
[1] Monash Univ, Clayton, Vic, Australia
基金
澳大利亚研究理事会;
关键词
D O I
10.1145/3524842.3528452
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Software vulnerabilities are prevalent in software systems, causing a variety of problems including deadlock, information loss, or system failures. Thus, early predictions of software vulnerabilities are critically important in safety-critical software systems. Various ML/DL-based approaches have been proposed to predict vulnerabilities at the file/function/method level. Recently, IVDetect (a graph-based neural network) is proposed to predict vulnerabilities at the function level. Yet, the IVDetect approach is still inaccurate and coarse-grained. In this paper, we propose LINEVUL, a Transformer-based line-level vulnerability prediction approach in order to address several limitations of the state-of-the-art IVDetect approach. Through an empirical evaluation of a large-scale real-world dataset with 188k+ C/C++ functions, we show that LINEVUL achieves (1) 160%-379% higher F1-measure for function-level predictions; (2) 12%-25% higher Top-10 Accuracy for line-level predictions; and (3) 29%-53% less Effort@20%Recall than the baseline approaches, highlighting the significant advancement of LINEVUL towards more accurate and more cost-effective line-level vulnerability predictions. Our additional analysis also shows that our LINEVUL is also very accurate (75%-100%) for predicting vulnerable functions affected by the Top-25 most dangerous CWEs, highlighting the potential impact of our LINEVUL in real-world usage scenarios.
引用
收藏
页码:608 / 620
页数:13
相关论文
共 50 条
  • [31] Towards the Use of Domain Knowledge to Enhance Transformer-Based Vulnerability Detection
    Marchetto, Alessandro
    Foulefack, Rosmael Zidane Lekeufack
    QUALITY OF INFORMATION AND COMMUNICATIONS TECHNOLOGY, QUATIC 2024, 2024, 2178 : 373 - 390
  • [32] Line-Level Incremental reSynthesis Techniques for FPGAs
    Chen, Doris
    Singh, Deshanand
    FPGA 11: PROCEEDINGS OF THE 2011 ACM/SIGDA INTERNATIONAL SYMPOSIUM ON FIELD PROGRAMMABLE GATE ARRAYS, 2011, : 133 - 142
  • [33] Transformer-based statement level vulnerability detection by cross-modal fine-grained features capture
    Tao, Wenxin
    Su, Xiaohong
    Ke, Yekun
    Han, Yi
    Zheng, Yu
    Wei, Hongwei
    KNOWLEDGE-BASED SYSTEMS, 2025, 316
  • [34] MPformer: A Transformer-Based Model for Earthen Ruins Climate Prediction
    Xu, Guodong
    Wang, Hai
    Ji, Shuo
    Ma, Yuhui
    Feng, Yi
    TSINGHUA SCIENCE AND TECHNOLOGY, 2024, 29 (06): : 1829 - 1838
  • [35] A Transformer-Based Model for the Prediction of Human Gaze Behavior on Videos
    Ozdel, Suleyman
    Rong, Yao
    Albaba, Berat Mert
    Kuo, Yen-Ling
    Wang, Xi
    Kasneci, Enkelejda
    PROCEEDINGS OF THE 2024 ACM SYMPOSIUM ON EYE TRACKING RESEARCH & APPLICATIONS, ETRA 2024, 2024,
  • [36] Transformer-Based Deep Learning Method for the Prediction of Ventilator Pressure
    Fan, Ruizhe
    2022 IEEE 2ND INTERNATIONAL CONFERENCE ON INFORMATION COMMUNICATION AND SOFTWARE ENGINEERING (ICICSE 2022), 2022, : 25 - 28
  • [37] Transformer-Based Architecture for Judgment Prediction and Explanation in Legal Proceedings
    Maqsood, Arooba
    Ul-Hasan, Adnan
    Shafait, Faisal
    DOCUMENT ANALYSIS SYSTEMS, DAS 2024, 2024, 14994 : 20 - 36
  • [38] Meta-learning for transformer-based prediction of potent compounds
    Hengwei Chen
    Jürgen Bajorath
    Scientific Reports, 13
  • [39] Molecular Descriptors Property Prediction Using Transformer-Based Approach
    Tran, Tuan
    Ekenna, Chinwe
    INTERNATIONAL JOURNAL OF MOLECULAR SCIENCES, 2023, 24 (15)
  • [40] Meta-learning for transformer-based prediction of potent compounds
    Chen, Hengwei
    Bajorath, Juergen
    SCIENTIFIC REPORTS, 2023, 13 (01)