Generative AI for pentesting: the good, the bad, the ugly

被引:1
|
作者
Hilario, Eric [1 ]
Azam, Sami [1 ]
Sundaram, Jawahar [2 ]
Mohammed, Khwaja Imran [1 ]
Shanmugam, Bharanidharan [1 ]
机构
[1] Charles Darwin Univ, Fac Sci & Technol, Energy & Resources Inst, Darwin, Australia
[2] Christ Acad Inst Adv Studies, Bangalore 560083, India
关键词
Cyber security; Generative AI; Large language models; Penetration testing; ChatGPT; 3.5;
D O I
10.1007/s10207-024-00835-x
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper examines the role of Generative AI (GenAI) and Large Language Models (LLMs) in penetration testing exploring the benefits, challenges, and risks associated with cyber security applications. Through the use of generative artificial intelligence, penetration testing becomes more creative, test environments are customised, and continuous learning and adaptation is achieved. We examined how GenAI (ChatGPT 3.5) helps penetration testers with options and suggestions during the five stages of penetration testing. The effectiveness of the GenAI tool was tested using a publicly available vulnerable machine from VulnHub. It was amazing how quickly they responded at each stage and provided better pentesting report. In this article, we discuss potential risks, unintended consequences, and uncontrolled AI development associated with pentesting.
引用
收藏
页码:2075 / 2097
页数:23
相关论文
共 50 条