MSFuzz: Augmenting Protocol Fuzzing with Message Syntax Comprehension via Large Language Models

被引:1
|
作者
Cheng, Mingjie [1 ,2 ]
Zhu, Kailong [1 ,2 ]
Chen, Yuanchao [1 ,2 ]
Yang, Guozheng [1 ,2 ]
Lu, Yuliang [1 ,2 ]
Lu, Canju [1 ,2 ]
机构
[1] Natl Univ Def Technol, Coll Elect Engn, Hefei 230037, Peoples R China
[2] Anhui Prov Key Lab Cyberspace Secur Situat Awarene, Hefei 230037, Peoples R China
关键词
fuzzing; syntax aware; protocol implementations; large language models; FUZZER;
D O I
10.3390/electronics13132632
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Network protocol implementations, as integral components of information communication, are critically important for security. Due to its efficiency and automation, fuzzing has become a popular method for protocol security detection. However, the existing protocol-fuzzing techniques face the critical problem of generating high-quality inputs. To address the problem, in this paper, we propose MSFuzz, which is a protocol-fuzzing method with message syntax comprehension. The core observation of MSFuzz is that the source code of protocol implementations contains detailed and comprehensive knowledge of the message syntax. Specifically, we leveraged the code-understanding capabilities of large language models to extract the message syntax from the source code and construct message syntax trees. Then, using these syntax trees, we expanded the initial seed corpus and designed a novel syntax-aware mutation strategy to guide the fuzzing. To evaluate the performance of MSFuzz, we compared it with the state-of-the-art (SOTA) protocol fuzzers, namely, AFLNET and CHATAFL. Experimental results showed that compared with AFLNET and CHATAFL, MSFuzz achieved average improvements of 22.53% and 10.04% in the number of states, 60.62% and 19.52% improvements in the number of state transitions, and 29.30% and 23.13% improvements in branch coverage. Additionally, MSFuzz discovered more vulnerabilities than the SOTA fuzzers.
引用
收藏
页数:19
相关论文
共 50 条
  • [1] Large Language Models Are Zero-Shot Fuzzers: Fuzzing Deep-Learning Libraries via Large Language Models
    Deng, Yinlin
    Xia, Chunqiu Steven
    Peng, Haoran
    Yang, Chenyuan
    Zhan, Lingming
    PROCEEDINGS OF THE 32ND ACM SIGSOFT INTERNATIONAL SYMPOSIUM ON SOFTWARE TESTING AND ANALYSIS, ISSTA 2023, 2023, : 423 - 435
  • [2] AUGMENTING AUTOTELIC AGENTS WITH LARGE LANGUAGE MODELS
    Colas, Cedric
    Teodorescu, Laetitia
    Oudeyer, Pierre-Yves
    Yuan, Xingdi
    Cote, Marc-Alexandre
    CONFERENCE ON LIFELONG LEARNING AGENTS, VOL 232, 2023, 232 : 205 - 226
  • [3] Augmenting large language models with chemistry tools
    Bran, Andres M.
    Cox, Sam
    Schilter, Oliver
    Baldassari, Carlo
    White, Andrew D.
    Schwaller, Philippe
    NATURE MACHINE INTELLIGENCE, 2024, 6 (05) : 525 - 535
  • [4] Augmenting interpretable models with large language models during training
    Chandan Singh
    Armin Askari
    Rich Caruana
    Jianfeng Gao
    Nature Communications, 14
  • [5] Augmenting interpretable models with large language models during training
    Singh, Chandan
    Askari, Armin
    Caruana, Rich
    Gao, Jianfeng
    NATURE COMMUNICATIONS, 2023, 14 (01)
  • [6] Skin and Syntax: Large Language Models in Dermatopathology
    Shah, Asghar
    Wahood, Samer
    Guermazi, Dorra
    Brem, Candice E.
    Saliba, Elie
    DERMATOPATHOLOGY, 2024, 11 (01): : 101 - 111
  • [7] Artificial intelligence: Augmenting telehealth with large language models
    Snoswell, Centaine L.
    Snoswell, Aaron J.
    Kelly, Jaimon T.
    Caffery, Liam J.
    Smith, Anthony C.
    JOURNAL OF TELEMEDICINE AND TELECARE, 2023,
  • [8] Psychological and Computational Models of Language Comprehension: In Defense of the Psychological Reality of Syntax
    Pereplyotchik, David
    CROATIAN JOURNAL OF PHILOSOPHY, 2011, 11 (31) : 31 - 72
  • [9] Fuzz4ALL: Universal Fuzzing with Large Language Models
    Xia, Chunqiu Steven
    Paltenghi, Matteo
    Tian, Jia Le
    Pradel, Michael
    Zhang, Lingming
    Proceedings - International Conference on Software Engineering, 2024, : 1547 - 1559
  • [10] WhiteFox: White-Box Compiler Fuzzing Empowered by Large Language Models
    Yang, Chenyuan
    Deng, Yinlin
    Lu, Runyu
    Yao, Jiayi
    Liu, Jiawei
    Jabbarvand, Reyhaneh
    Zhang, Lingming
    Proceedings of the ACM on Programming Languages, 2024, 8 (OOPSLA2)