Maritime platform defense with deep reinforcement learning

被引:2
|
作者
Markowitz, Jared [1 ]
Sheffield, Ryan [1 ]
Mullins, Galen [1 ]
机构
[1] Johns Hopkins Univ, Appl Phys Lab, Johns Hopkins Rd, Laurel, MD 20707 USA
关键词
Deep Reinforcement Learning; Maritime Platform Defense; AI Safety; Continual Learning;
D O I
10.1117/12.2618808
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present a method for applying deep reinforcement learning to maritime platform defense, showing how to successfully train agents to schedule countermeasures for defending a fleet of ships against stochastic raids in a simulated environment. Our Schedule Evaluation Simulation (SEvSim) environment was developed using extensive input from subject matter experts and contains realistic threat characteristics, weapon efficacies, and constraints among weapons. Our approach includes novelty in both the representation of the system state and the neural network architecture: threats are represented as vectors containing information on the projected effect of different scheduling actions on their viability and fed to network input "slots" in randomized locations. Agents are trained using Proximal Policy Optimization, a state-of-the-art method for model-free learning. We evaluate the performance of our approach, finding that it learns scheduling strategies that both reliably neutralize threats and conserve inventory. We subsequently discuss the remaining challenges involved in bringing neural-network-based control to realization in this application space. Among these challenges are the needs to integrate humans into the loop, provide safety assurances, and enable continual learning.
引用
收藏
页数:7
相关论文
共 50 条
  • [21] Security defense strategy algorithm for Internet of Things based on deep reinforcement learning
    Feng, Xuecai
    Han, Jikai
    Zhang, Rui
    Xu, Shuo
    Xia, Hui
    [J]. HIGH-CONFIDENCE COMPUTING, 2024, 4 (01):
  • [22] Instance-based defense against adversarial attacks in Deep Reinforcement Learning
    Garcia, Javier
    Sagredo, Ismael
    [J]. ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2022, 107
  • [23] The Advance of Reinforcement Learning and Deep Reinforcement Learning
    Lyu, Le
    Shen, Yang
    Zhang, Sicheng
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ELECTRICAL ENGINEERING, BIG DATA AND ALGORITHMS (EEBDA), 2022, : 644 - 648
  • [24] Deep Reinforcement Learning Based Mimicry Defense System for IoT Message Transmission
    Wang, Zhihao
    Jiang, Dingde
    Chen, Jianguang
    Yang, Wei
    [J]. SIMULATION TOOLS AND TECHNIQUES, SIMUTOOLS 2021, 2022, 424 : 421 - 431
  • [25] Intelligent defense strategies: Comprehensive attack detection in VANET with deep reinforcement learning
    Sultana, Rukhsar
    Grover, Jyoti
    Tripathi, Meenakshi
    [J]. PERVASIVE AND MOBILE COMPUTING, 2024, 103
  • [26] Research on multidimensional dynamic defense strategy for microservice based on deep reinforcement learning
    Zhou, Dacheng
    Chen, Hongchang
    He, Weizhen
    Cheng, Guozhen
    Hu, Hongchao
    [J]. Tongxin Xuebao/Journal on Communications, 2023, 44 (04): : 50 - 63
  • [27] Defense Strategies Against Adversarial Jamming Attacks via Deep Reinforcement Learning
    Wang, Feng
    Zhong, Chen
    Gursoy, M. Cenk
    Velipasalar, Senem
    [J]. 2020 54TH ANNUAL CONFERENCE ON INFORMATION SCIENCES AND SYSTEMS (CISS), 2020, : 336 - 341
  • [28] Three-Tier Computing Platform Optimization: A Deep Reinforcement Learning Approach
    Chidume, Chidiebere Sunday
    Okopi, Solomon Inalegwu
    Sesay, Taiwo
    Materu, Irene Simon
    Asenso, Theophilus Quachie
    [J]. MOBILE INFORMATION SYSTEMS, 2022, 2022
  • [29] Deep Reinforcement Learning with Corrective Feedback for Autonomous UAV Landing on a Mobile Platform
    Wu, Lizhen
    Wang, Chang
    Zhang, Pengpeng
    Wei, Changyun
    [J]. DRONES, 2022, 6 (09)
  • [30] Parallel Gym Gazebo: a Scalable Parallel Robot Deep Reinforcement Learning Platform
    Liang, Zhen
    Cai, Zhongxuan
    Li, Minglong
    Yang, Wenjing
    [J]. 2019 IEEE 31ST INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2019), 2019, : 206 - 213