Smart Security Audit: Reinforcement Learning with a Deep Neural Network Approximator

被引:14
|
作者
Pozdniakov, Konstantin [1 ]
Alonso, Eduardo [1 ]
Stankovic, Vladimir [1 ]
Tam, Kimberly [2 ]
Jones, Kevin [2 ]
机构
[1] City Univ London, London, England
[2] Univ Plymouth, Plymouth, Devon, England
关键词
Pentesting; audit; Q-learning; reinforcement learning; deep neural network; MODEL CHECKING;
D O I
10.1109/cybersa49311.2020.9139683
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
A significant challenge in modern computer security is the growing skill gap as intruder capabilities increase, making it necessary to begin automating elements of penetration testing so analysts can contend with the growing number of cyber threats. In this paper, we attempt to assist human analysts by automating a single host penetration attack. To do so, a smart agent performs different attack sequences to find vulnerabilities in a target system. As it does so, it accumulates knowledge, learns new attack sequences and improves its own internal penetration testing logic. As a result, this agent (AgentPen for simplicity) is able to successfully penetrate hosts it has never interacted with before. A computer security administrator using this tool would receive a comprehensive, automated sequence of actions leading to a security breach, highlighting potential vulnerabilities, and reducing the amount of menial tasks a typical penetration tester would need to execute. To achieve autonomy, we apply an unsupervised machine learning algorithm, Q-learning, with an approximator that incorporates a deep neural network architecture. The security audit itself is modelled as a Markov Decision Process in order to test a number of decision-making strategies and compare their convergence to optimality. A series of experimental results is presented to show how this approach can be effectively used to automate penetration testing using a scalable, i.e. not exhaustive, and adaptive approach.
引用
收藏
页数:8
相关论文
共 50 条
  • [21] Untying cable by combining 3D deep neural network with deep reinforcement learning
    Fan, Zheming
    Shao, Wanpeng
    Hayashi, Toyohiro
    Ohashi, Takeshi
    [J]. ADVANCED ROBOTICS, 2023, 37 (05) : 380 - 394
  • [22] Network Planning with Deep Reinforcement Learning
    Zhu, Hang
    Gupta, Varun
    Ahuja, Satyajeet Singh
    Tian, Yuandong
    Zhang, Ying
    Jin, Xin
    [J]. SIGCOMM '21: PROCEEDINGS OF THE 2021 ACM SIGCOMM 2021 CONFERENCE, 2021, : 258 - 271
  • [23] Neural Network Ensembles in Reinforcement Learning
    Stefan Faußer
    Friedhelm Schwenker
    [J]. Neural Processing Letters, 2015, 41 : 55 - 69
  • [24] Neural Network Ensembles in Reinforcement Learning
    Fausser, Stefan
    Schwenker, Friedhelm
    [J]. NEURAL PROCESSING LETTERS, 2015, 41 (01) : 55 - 69
  • [25] Deep Convolutional Neural Network Assisted Reinforcement Learning Based Mobile Network Power Saving
    Wu, Shangbin
    Wang, Yue
    Bai, Lu
    [J]. IEEE ACCESS, 2020, 8 : 93671 - 93681
  • [26] Reinforcement learning using a grid based function approximator
    Sung, A
    Merke, A
    Riedmiller, M
    [J]. BIOMIMETIC NEURAL LEARNING FOR INTELLIGENT ROBOTS: INTELLIGENT SYSTEMS, COGNITIVE ROBOTICS, AND NEUROSCIENCE, 2005, 3575 : 235 - 244
  • [27] Reducing memory requirements of scope approximator in reinforcement learning
    Michalski, A
    [J]. INTELLIGENT INFORMATION SYSTEMS 2002, PROCEEDINGS, 2002, 17 : 341 - 350
  • [28] REINFORCEMENT LEARNING WITH SAFE EXPLORATION FOR NETWORK SECURITY
    Dai, Canhuang
    Xiao, Liang
    Wan, Xiaoyue
    Chen, Ye
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 3057 - 3061
  • [29] Tree-like function approximator in reinforcement learning
    Hwang, Kao-Shing
    Chen, Yu-Jen
    [J]. IECON 2007: 33RD ANNUAL CONFERENCE OF THE IEEE INDUSTRIAL ELECTRONICS SOCIETY, VOLS 1-3, CONFERENCE PROCEEDINGS, 2007, : 904 - 907
  • [30] Shallow Network Training With Dynamic Sample Weights Decay - a Potential Function Approximator for Reinforcement Learning
    Ghignone, Leo
    Barlow, Michael
    [J]. 2019 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI 2019), 2019, : 149 - 154