Few-VulD: A Few-shot learning framework for software vulnerability detection☆ ☆

被引:0
|
作者
Zheng, Tianming [1 ]
Liu, Haojun [2 ]
Xu, Hang [1 ]
Chen, Xiang [1 ]
Yi, Ping [1 ]
Wu, Yue [1 ]
机构
[1] Shanghai Jiao Tong Univ, Sch Elect Informat & Elect Engn, Shanghai, Peoples R China
[2] Carnegie Mellon Univ, Sch Comp Sci, Pittsburgh, PA USA
基金
国家重点研发计划;
关键词
Vulnerability detection; Few-shot learning; Meta-learning; BiLSTM; Artificial intelligence; Deep learning; NEURAL-NETWORKS;
D O I
10.1016/j.cose.2024.103992
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The rapid development of artificial intelligence (AI) has led to the introduction of numerous software vulnerability detection methods based on deep learning algorithms. However, a significant challenge is their dependency on large volumes of code samples for effective training. This requirement poses a considerable hurdle, particularly when adapting to diverse software application scenarios and various vulnerability types, where gathering sufficient and relevant training data for different classification tasks is often arduous. To address the challenge, this paper introduces Few-VulD, a novel framework for software vulnerability detection based on few-shot learning. This framework is designed to be efficiently trained with a minimal number of samples from a variety of existing classification tasks. Its key advantage lies in its ability to rapidly adapt to new vulnerability detection tasks, such as identifying new types of vulnerabilities, with only a small set of learning samples. This capability is particularly beneficial in scenarios where available vulnerability samples are limited. We compare Few-VulD with five state-of-the-art methods on the SySeVR and Big-Vul datasets. On the SySeVR dataset, Few-VulD outperforms all other methods, achieving a recall rate of 87.9% and showing an improvement of 11.7% to 57.8%. On the Big-Vul dataset, Few-VulD outperforms three of the methods, including one that utilizes a pretrained large language model (LLM), with recall improvements ranging from 8.5% to 40.1%. The other two methods employ pretrained LLMs from Microsoft CodeXGLUE (Lu et al., 2021). Few-VulD reaches 78.7% and 95.5% of their recall rates without the need for extensive data pretraining. The performance proves the effectiveness of Few-VulD in vulnerability detection tasks with limited samples.
引用
收藏
页数:13
相关论文
共 50 条
  • [31] Fast Hierarchical Learning for Few-Shot Object Detection
    She, Yihang
    Bhat, Goutam
    Danelljan, Martin
    Yu, Fisher
    2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 1993 - 2000
  • [32] Few-shot learning for signal detection in wideband spectrograms
    Li, Weihao
    Deng, Wen
    Wang, Keren
    You, Ling
    Huang, Zhitao
    DIGITAL SIGNAL PROCESSING, 2025, 162
  • [33] Few-Shot Anomaly Detection in Text with Deviation Learning
    Das, Anindya Sundar
    Ajay, Aravind
    Saha, Sriparna
    Bhuyan, Monowar
    NEURAL INFORMATION PROCESSING, ICONIP 2023, PT II, 2024, 14448 : 425 - 438
  • [34] A Gated Few-shot Learning Model For Anomaly Detection
    Huang, Shaohan
    Liu, Yi
    Fung, Carol
    An, Wanhe
    He, Rong
    Zhao, Yining
    Yang, Hailong
    Luan, Zhongzhi
    2020 34TH INTERNATIONAL CONFERENCE ON INFORMATION NETWORKING (ICOIN 2020), 2020, : 505 - 509
  • [35] Extensively Matching for Few-shot Learning Event Detection
    Viet Dac Lai
    Dernoncourt, Franck
    Thien Huu Nguyen
    NARRATIVE UNDERSTANDING, STORYLINES, AND EVENTS, 2020, : 38 - 45
  • [36] A few-shot learning framework for planar pushing of unknown objects
    Ziyan Gao
    Armagan Elibol
    Nak Young Chong
    Intelligent Service Robotics, 2022, 15 : 335 - 350
  • [37] ContrastNet: A Contrastive Learning Framework for Few-Shot Text Classification
    Chen, Junfan
    Zhang, Richong
    Mao, Yongyi
    Xu, Jie
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 10492 - 10500
  • [38] FedFSLAR: A Federated Learning Framework for Few-shot Action Recognition
    Nguyen Anh Tu
    Abu, Assanali
    Aikyn, Nartay
    Makhanov, Nursultan
    Lee, Min-Ho
    Khiem Le-Huy
    Wong, Kok-Seng
    2024 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION WORKSHOPS, WACVW 2024, 2024, : 270 - 279
  • [39] Sylph: A Hypernetwork Framework for Incremental Few-shot Object Detection
    Yin, Li
    Perez-Rua, Juan M.
    Liang, Kevin J.
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 9025 - 9035
  • [40] A few-shot learning framework for planar pushing of unknown objects
    Gao, Ziyan
    Elibol, Armagan
    Chong, Nak Young
    INTELLIGENT SERVICE ROBOTICS, 2022, 15 (03) : 335 - 350