Utilizing Fine-Tuning of Large Language Models for Generating Synthetic Payloads: Enhancing Web Application Cybersecurity through Innovative Penetration Testing Techniques

被引:0
|
作者
Cirkovic, Stefan [1 ]
Mladenovic, Vladimir [1 ]
Tomic, Sinisa [2 ]
Drljaca, Dalibor [2 ]
Ristic, Olga [1 ]
机构
[1] Univ Kragujevac, Fac Tech Sci, Cacak 32000, Serbia
[2] Pan European Univ Apeiron, Fac Informat Technol, Banja Luka 78101, Bosnia & Herceg
来源
CMC-COMPUTERS MATERIALS & CONTINUA | 2025年 / 82卷 / 03期
关键词
LLM; GPT-2; XSS; SQL injection; command injection; evaluation loss perplexity;
D O I
10.32604/cmc.2025.059696
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
With the increasing use of web applications, challenges in the field of cybersecurity are becoming more complex. This paper explores the application of fine-tuned large language models (LLMs) for the automatic generation of synthetic attacks, including XSS (Cross-Site Scripting), SQL Injections, and Command Injections. A web application has been developed that allows penetration testers to quickly generate high-quality payloads without the need for in-depth knowledge of artificial intelligence. The fine-tuned language model demonstrates the capability to produce synthetic payloads that closely resemble real-world attacks. This approach not only improves the model's precision and dependability but also serves as a practical resource for cybersecurity professionals to enhance the security of web applications. The methodology and structured implementation underscore the importance and potential of advanced language models in cybersecurity, illustrating their effectiveness in generating high-quality synthetic data for penetration testing purposes. The research results demonstrate that this approach enables the identification of vulnerabilities that traditional methods may not uncover, providing deeper insights into potential threats and enhancing overall security measures. The performance evaluation of the model indicated satisfactory results, while further hyperparameter optimization could improve accuracy and generalization capabilities. This research represents a significant step forward in improving web application security and opens new opportunities for the use of LLMs in security testing, thereby contributing to the development of more effective cybersecurity strategies.
引用
收藏
页码:4409 / 4430
页数:22
相关论文
共 9 条
  • [1] HackMentor: Fine-Tuning Large Language Models for Cybersecurity
    Zhang, Jie
    Wen, Hui
    Deng, Liting
    Xin, Mingfeng
    Li, Zhi
    Li, Lun
    Zhu, Hongsong
    Sun, Limin
    2023 IEEE 22ND INTERNATIONAL CONFERENCE ON TRUST, SECURITY AND PRIVACY IN COMPUTING AND COMMUNICATIONS, TRUSTCOM, BIGDATASE, CSE, EUC, ISCI 2023, 2024, : 452 - 461
  • [2] Personalized Large Language Models through Parameter Efficient Fine-Tuning Techniques
    Braga, Marco
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 3076 - 3076
  • [3] Enhancing Chinese Essay Discourse Logic Evaluation Through Optimized Fine-Tuning of Large Language Models
    Song, Jinwang
    Song, Yanxin
    Zhou, Guangyu
    Fu, Wenhui
    Zhang, Kunli
    Zan, Hongying
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, PT V, NLPCC 2024, 2025, 15363 : 342 - 352
  • [4] SDD-LawLLM: Advancing Intelligent Legal Systems Through Synthetic Data-Driven Fine-Tuning of Large Language Models
    Ma, Hanjie
    Lu, Yuhang
    Xiao, Zhengdong
    Feng, Jie
    Zhang, Haixiang
    Yu, Jian
    ELECTRONICS, 2025, 14 (04):
  • [5] Enhancing Chinese comprehension and reasoning for large language models: an efficient LoRA fine-tuning and tree of thoughts framework
    Chen, Songlin
    Wang, Weicheng
    Chen, Xiaoliang
    Zhang, Maolin
    Lu, Peng
    Li, Xianyong
    Du, Yajun
    JOURNAL OF SUPERCOMPUTING, 2025, 81 (01):
  • [6] Enhancing the security of edge-AI runtime environments: a fine-tuning method based on large language models
    Tang, Di
    Xiao, Peng
    Zheng, Tao
    Li, Xiang
    Yang, Cuibo
    WIRELESS NETWORKS, 2025, 31 (02) : 1825 - 1838
  • [7] OptimalMEE: Optimizing Large Language Models for Medical Event Extraction Through Fine-Tuning and Post-hoc Verification
    Sun, Yaoqian
    Wu, Dan
    Chen, Zikang
    Cai, Hailing
    An, Jiye
    ARTIFICIAL INTELLIGENCE IN MEDICINE, PT I, AIME 2024, 2024, 14844 : 303 - 311
  • [8] Enhancing Health Mention Classification Through Reexamining Misclassified Samples and Robust Fine-Tuning Pre-Trained Language Models
    Meng, Deyu
    Phuntsho, Tshewang
    Gonsalves, Tad
    IEEE ACCESS, 2024, 12 : 190445 - 190453
  • [9] LLaMA-Reviewer: Advancing Code Review Automation with Large Language Models through Parameter-Efficient Fine-Tuning
    Lu, Junyi
    Yu, Lei
    Li, Xiaojia
    Yang, Li
    Zuo, Chun
    2023 IEEE 34TH INTERNATIONAL SYMPOSIUM ON SOFTWARE RELIABILITY ENGINEERING, ISSRE, 2023, : 647 - 658