MSFuzz: Augmenting Protocol Fuzzing with Message Syntax Comprehension via Large Language Models

被引:1
|
作者
Cheng, Mingjie [1 ,2 ]
Zhu, Kailong [1 ,2 ]
Chen, Yuanchao [1 ,2 ]
Yang, Guozheng [1 ,2 ]
Lu, Yuliang [1 ,2 ]
Lu, Canju [1 ,2 ]
机构
[1] Natl Univ Def Technol, Coll Elect Engn, Hefei 230037, Peoples R China
[2] Anhui Prov Key Lab Cyberspace Secur Situat Awarene, Hefei 230037, Peoples R China
关键词
fuzzing; syntax aware; protocol implementations; large language models; FUZZER;
D O I
10.3390/electronics13132632
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Network protocol implementations, as integral components of information communication, are critically important for security. Due to its efficiency and automation, fuzzing has become a popular method for protocol security detection. However, the existing protocol-fuzzing techniques face the critical problem of generating high-quality inputs. To address the problem, in this paper, we propose MSFuzz, which is a protocol-fuzzing method with message syntax comprehension. The core observation of MSFuzz is that the source code of protocol implementations contains detailed and comprehensive knowledge of the message syntax. Specifically, we leveraged the code-understanding capabilities of large language models to extract the message syntax from the source code and construct message syntax trees. Then, using these syntax trees, we expanded the initial seed corpus and designed a novel syntax-aware mutation strategy to guide the fuzzing. To evaluate the performance of MSFuzz, we compared it with the state-of-the-art (SOTA) protocol fuzzers, namely, AFLNET and CHATAFL. Experimental results showed that compared with AFLNET and CHATAFL, MSFuzz achieved average improvements of 22.53% and 10.04% in the number of states, 60.62% and 19.52% improvements in the number of state transitions, and 29.30% and 23.13% improvements in branch coverage. Additionally, MSFuzz discovered more vulnerabilities than the SOTA fuzzers.
引用
收藏
页数:19
相关论文
共 50 条
  • [31] Reducing hallucinations of large language models via hierarchical semantic piece
    Yanyi Liu
    Qingwen Yang
    Jiawei Tang
    Tiezheng Guo
    Chen Wang
    Pan Li
    Sai Xu
    Xianlin Gao
    Zhi Li
    Jun Liu
    Yingyou Wen
    Complex & Intelligent Systems, 2025, 11 (5)
  • [32] MONITORASSISTANT: Simplifying Cloud Service Monitoring via Large Language Models
    Yu, Zhaoyang
    Ma, Minghua
    Zhang, Chaoyun
    Qin, Si
    Kang, Yu
    Bansal, Chetan
    Rajmohan, Saravan
    Dang, Yingnong
    Pei, Changhua
    Pei, Dan
    Lin, Qingwei
    Zhang, Dongmei
    COMPANION PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON THE FOUNDATIONS OF SOFTWARE ENGINEERING, FSE COMPANION 2024, 2024, : 38 - 49
  • [33] Fast Approximate Inference for Arbitrarily Large Semiparametric Regression Models via Message Passing
    Wand, M. P.
    JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2017, 112 (517) : 137 - 156
  • [34] Enhanced Story Comprehension for Large Language Models through Dynamic Document-Based Knowledge Graphs
    Andrus, Berkeley R.
    Nasiri, Yeganeh
    Cui, Shilong
    Cullen, Benjamin
    Fulda, Nancy
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 10436 - 10444
  • [35] Fluid Transformers and Creative Analogies: Exploring Large Language Models' Capacity for Augmenting Cross-Domain Analogical Creativity
    Ding, Zijian
    Srinivasan, Arvind
    MacNeil, Stephen
    Chan, Joel
    2023 PROCEEDINGS OF THE 15TH CONFERENCE ON CREATIVITY AND COGNITION, C&C 2023, 2023, : 489 - 505
  • [36] SpecGen: Automated Generation of Formal Program Specifications via Large Language Models
    Ma, Lezhi
    Liu, Shangqing
    Li, Yi
    Xie, Xiaofei
    Bu, Lei
    arXiv, 1600,
  • [37] Automatic Bug Fixing via Deliberate Problem Solving with Large Language Models
    Weng, Guoyang
    Andrzejak, Artur
    2023 IEEE 34TH INTERNATIONAL SYMPOSIUM ON SOFTWARE RELIABILITY ENGINEERING WORKSHOPS, ISSREW, 2023, : 34 - 36
  • [38] Correctable Landmark Discovery Via Large Models for Vision-Language Navigation
    Lin B.
    Nie Y.
    Wei Z.
    Zhu Y.
    Xu H.
    Ma S.
    Liu J.
    Liang X.
    IEEE Transactions on Pattern Analysis and Machine Intelligence, 2024, 46 (12) : 1 - 14
  • [39] Improving Patch Correctness Analysis via Random Testing and Large Language Models
    Molina, Facundo
    Manuel Copia, Juan
    Gorla, Alessandra
    2024 IEEE CONFERENCE ON SOFTWARE TESTING, VERIFICATION AND VALIDATION, ICST 2024, 2024, : 317 - 328
  • [40] Enhancing Financial Sentiment Analysis via Retrieval Augmented Large Language Models
    Zhang, Boyu
    Yang, Hongyang
    Zhou, Tianyu
    Babar, Ali
    Liu, Xiao-Yang
    PROCEEDINGS OF THE 4TH ACM INTERNATIONAL CONFERENCE ON AI IN FINANCE, ICAIF 2023, 2023, : 349 - 356