Military Decision Support with Actor and Critic Reinforcement Learning Agents

被引:1
|
作者
Ma, Jungmok [1 ]
机构
[1] Korea Natl Def Univ, Dept Natl Def Sci, Nonsan, South Korea
关键词
Reinforcement learning; Military decision support; Actor and critic; Weapon selection; Battle damage assessment; INTRUSION DETECTION; UAV; AUTHENTICATION; DEFENSE;
D O I
10.14429/dsj.74.18864
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
While the recent advanced military operational concept requires intelligent support of command and control, Reinforcement Learning (RL) has not been actively studied in the military domain. This study points out the limitations of RL for military applications from a literature review and aims to improve the understanding of RL for military decision support under these limitations. Most of all, the black box characteristic of Deep RL makes the internal process difficult to understand, in addition to the complex simulation tools. A scalable weapon selection RL framework is built, which can be solved either by a tabular form or a neural network form. The transition of the Deep Q -Network (DQN) solution to the tabular form allows for effective comparison of the results to the Q -learning solution. Furthermore, rather than using one or two RL models selectively as before, RL models are divided into an actor and a critic, and systematically compared. A random agent, Q -learning and DQN agents as critics, a Policy Gradient (PG) agent as an actor, Trust Region Policy Optimization (TRPO) and Proximal Policy Optimization (PPO) agents as an actor -critic approach are designed, trained, and tested. The performance results show that the trained DQN and PPO agents are the best decision support candidates for the weapon selection RL framework.
引用
收藏
页码:389 / 398
页数:10
相关论文
共 50 条
  • [1] Deep Reinforcement Learning in VizDoom via DQN and Actor-Critic Agents
    Bakhanova, Maria
    Makarov, Ilya
    ADVANCES IN COMPUTATIONAL INTELLIGENCE, IWANN 2021, PT I, 2021, 12861 : 138 - 150
  • [2] A World Model for Actor–Critic in Reinforcement Learning
    A. I. Panov
    L. A. Ugadiarov
    Pattern Recognition and Image Analysis, 2023, 33 : 467 - 477
  • [3] Selector-Actor-Critic and Tuner-Actor-Critic Algorithms for Reinforcement Learning
    Masadeh, Ala'eddin
    Wang, Zhengdao
    Kamal, Ahmed E.
    2019 11TH INTERNATIONAL CONFERENCE ON WIRELESS COMMUNICATIONS AND SIGNAL PROCESSING (WCSP), 2019,
  • [4] Multi-actor mechanism for actor-critic reinforcement learning
    Li, Lin
    Li, Yuze
    Wei, Wei
    Zhang, Yujia
    Liang, Jiye
    INFORMATION SCIENCES, 2023, 647
  • [5] A deep actor critic reinforcement learning framework for learning to rank
    Padhye, Vaibhav
    Lakshmanan, Kailasam
    NEUROCOMPUTING, 2023, 547
  • [6] Actor-Critic reinforcement learning for optimal design of piping support constraint combinations
    Ham, Jong-Ho
    An, Jung-Eun
    Lee, Hee-Sung
    Park, Gun-il
    Lee, Dong-Yeon
    INTERNATIONAL JOURNAL OF NAVAL ARCHITECTURE AND OCEAN ENGINEERING, 2022, 14
  • [7] A World Model for Actor-Critic in Reinforcement Learning
    Panov, A. I.
    Ugadiarov, L. A.
    PATTERN RECOGNITION AND IMAGE ANALYSIS, 2023, 33 (03) : 467 - 477
  • [8] Curious Hierarchical Actor-Critic Reinforcement Learning
    Roeder, Frank
    Eppe, Manfred
    Nguyen, Phuong D. H.
    Wermter, Stefan
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2020, PT II, 2020, 12397 : 408 - 419
  • [9] Actor-Critic based Improper Reinforcement Learning
    Zaki, Mohammadi
    Mohan, Avinash
    Gopalan, Aditya
    Mannor, Shie
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [10] The Actor-Dueling-Critic Method for Reinforcement Learning
    Wu, Menghao
    Gao, Yanbin
    Jung, Alexander
    Zhang, Qiang
    Du, Shitong
    SENSORS, 2019, 19 (07)