Dynamic Scheduling of Cybersecurity Analysts for Minimizing Risk Using Reinforcement Learning

被引:38
|
作者
Ganesan, Rajesh [1 ]
Jajodia, Sushil [2 ]
Shah, Ankit [2 ]
Cam, Hasan [3 ]
机构
[1] George Mason Univ, Dept Syst Engn & Operat Res, Mail Stop 4A6, Fairfax, VA 22030 USA
[2] George Mason Univ, Ctr Secure Informat Syst, Mail Stop 5B5, Fairfax, VA 22030 USA
[3] Army Res Lab, 2800 Powder Mill Rd, Adelphi, MD 20783 USA
基金
美国国家科学基金会;
关键词
Cybersecurity; Analysts; Dynamic Scheduling; Cybersecurity analysts; dynamic scheduling; genetic algorithm; integer programming; optimization; reinforcement learning; resource allocation; risk mitigation;
D O I
10.1145/2882969
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
An important component of the cyber-defense mechanism is the adequate staffing levels of its cybersecurity analyst workforce and their optimal assignment to sensors for investigating the dynamic alert traffic. The ever-increasing cybersecurity threats faced by today's digital systems require a strong cyber-defense mechanism that is both reactive in its response to mitigate the known risk and proactive in being prepared for handling the unknown risks. In order to be proactive for handling the unknown risks, the above workforce must be scheduled dynamically so the system is adaptive to meet the day-to-day stochastic demands on its workforce (both size and expertise mix). The stochastic demands on the workforce stem from the varying alert generation and their significance rate, which causes an uncertainty for the cybersecurity analyst scheduler that is attempting to schedule analysts for work and allocate sensors to analysts. Sensor data are analyzed by automatic processing systems, and alerts are generated. A portion of these alerts is categorized to be significant, which requires thorough examination by a cybersecurity analyst. Risk, in this article, is defined as the percentage of significant alerts that are not thoroughly analyzed by analysts. In order to minimize risk, it is imperative that the cyber-defense system accurately estimates the future significant alert generation rate and dynamically schedules its workforce to meet the stochastic workload demand to analyze them. The article presents a reinforcement learning-based stochastic dynamic programming optimization model that incorporates the above estimates of future alert rates and responds by dynamically scheduling cybersecurity analysts to minimize risk (i.e., maximize significant alert coverage by analysts) and maintain the risk under a pre-determined upper bound. The article tests the dynamic optimization model and compares the results to an integer programming model that optimizes the static staffing needs based on a daily-average alert generation rate with no estimation of future alert rates (static workforce model). Results indicate that over a finite planning horizon, the learning-based optimization model, through a dynamic (on-call) workforce in addition to the static workforce, (a) is capable of balancing risk between days and reducing overall risk better than the static model, (b) is scalable and capable of identifying the quantity and the right mix of analyst expertise in an organization, and (c) is able to determine their dynamic (on-call) schedule and their sensor-to-analyst allocation in order to maintain risk below a given upper bound. Several meta-principles are presented, which are derived from the optimization model, and they further serve as guiding principles for hiring and scheduling cybersecurity analysts. Days-off scheduling was performed to determine analyst weekly work schedules that met the cybersecurity system's workforce constraints and requirements.
引用
收藏
页码:1 / 21
页数:21
相关论文
共 50 条
  • [21] Dynamic job-shop scheduling in smart manufacturing using deep reinforcement learning
    Wang, Libing
    Hu, Xin
    Wang, Yin
    Xu, Sujie
    Ma, Shijun
    Yang, Kexin
    Liu, Zhijun
    Wang, Weidong
    [J]. COMPUTER NETWORKS, 2021, 190 (190)
  • [22] Scheduling in Multiagent Systems Using Reinforcement Learning
    I. K. Minashina
    R. A. Gorbachev
    E. M. Zakharova
    [J]. Doklady Mathematics, 2022, 106 : S70 - S78
  • [23] Scheduling in Multiagent Systems Using Reinforcement Learning
    Minashina, I. K.
    Gorbachev, R. A.
    Zakharova, E. M.
    [J]. DOKLADY MATHEMATICS, 2022, 106 (SUPPL 1) : S70 - S78
  • [24] Solving a Dynamic Scheduling Problem for a Manufacturing System with Reinforcement Learning
    Heik, David
    Bahrpeyma, Fouad
    Reichelt, Dirk
    [J]. INTELLIGENT SYSTEMS AND APPLICATIONS, VOL 2, INTELLISYS 2023, 2024, 823 : 413 - 432
  • [25] Deep Reinforcement Learning for Dynamic Workflow Scheduling in Cloud Environment
    Dong, Tingting
    Xue, Fei
    Xiao, Changbai
    Zhang, Jiangjiang
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON SERVICES COMPUTING (SCC 2021), 2021, : 107 - 115
  • [26] Dynamic Job Shop Scheduling via Deep Reinforcement Learning
    Liang, Xinjie
    Song, Wen
    Wei, Pengfei
    [J]. 2023 IEEE 35TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, ICTAI, 2023, : 369 - 376
  • [27] Deep reinforcement learning for dynamic scheduling of a flexible job shop
    Liu, Renke
    Piplani, Rajesh
    Toro, Carlos
    [J]. INTERNATIONAL JOURNAL OF PRODUCTION RESEARCH, 2022, 60 (13) : 4049 - 4069
  • [28] READYS: A Reinforcement Learning Based Strategy for Heterogeneous Dynamic Scheduling
    Grinsztajn, Nathan
    Beaumont, Olivier
    Jeannot, Emmanuel
    Preux, Philippe
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON CLUSTER COMPUTING (CLUSTER 2021), 2021, : 70 - 81
  • [29] Reinforcement learning with dynamic convex risk measures
    Coache, Anthony
    Jaimungal, Sebastian
    [J]. MATHEMATICAL FINANCE, 2024, 34 (02) : 557 - 587
  • [30] Genetic Programming and Reinforcement Learning on Learning Heuristics for Dynamic Scheduling: A Preliminary Comparison
    Xu, Meng
    Mei, Yi
    Zhang, Fangfang
    Zhang, Mengjie
    [J]. IEEE COMPUTATIONAL INTELLIGENCE MAGAZINE, 2024, 19 (02) : 18 - 33