Missile Evasion Maneuver Generation with Model-free Deep Reinforcement Learning

被引:0
|
作者
Ozbek, Muhammed Murat [1 ]
Koyuncu, Emre [1 ]
机构
[1] Istanbul Tech Univ, Aerosp Res Ctr, AI Guidance Nav & Control Res Grp, Istanbul, Turkiye
关键词
Reinforcement Learning; UCAV; Control; UAV;
D O I
10.1109/RAST57548.2023.10197852
中图分类号
V [航空、航天];
学科分类号
08 ; 0825 ;
摘要
Unmanned Combat Aerial Vehicles (UCAVs) play a significant role in modern military conflicts as they can perform intelligence, surveillance, reconnaissance, and target acquisition missions while carrying aircraft ordnance like missiles and bombs, and Anti-Tank Guided Missiles (ATGMs). However, the increased use of UCAVs has also led to more advanced anti-UCAV solutions in air defense. The paper proposes a deep reinforcement learning approach for generating online missile-evading maneuvers for combat aerial vehicles. The problem is made complicated by the missile's 8 Mach speed and the aircraft's limited 2.5 Mach speed. The system employs Twin Delayed Deep Deterministic Policy Gradient(TD3), one of the most known deep reinforcement learning algorithms, to train an agent to make real-time decisions on the best evasion tactics in a complex combat environment. A two-term reward function is used, with sparse rewards at terminal states and continuous rewards through the geometry of the combat. Aileron, rudder, and elevator controls are given directly to the algorithm to ensure all potential escape maneuvers are visible. The proposed methodology achieved a 59% success rate in extensive simulations, demonstrating its potential to enhance aerial vehicles' combat capabilities.
引用
收藏
页数:6
相关论文
共 50 条
  • [1] Model-Free Deep Inverse Reinforcement Learning by Logistic Regression
    Eiji Uchibe
    [J]. Neural Processing Letters, 2018, 47 : 891 - 905
  • [2] Model-free Deep Reinforcement Learning for Urban Autonomous Driving
    Chen, Jianyu
    Yuan, Bodi
    Tomizuka, Masayoshi
    [J]. 2019 IEEE INTELLIGENT TRANSPORTATION SYSTEMS CONFERENCE (ITSC), 2019, : 2765 - 2771
  • [4] Deep Reinforcement Learning for Autonomous Model-Free Navigation with Partial Observability
    Tapia, Daniel
    Parras, Juan
    Zazo, Santiago
    [J]. 2019 27TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO), 2019,
  • [5] A Model-free Deep Reinforcement Learning Approach for Robotic Manipulators Path Planning
    Liu, Wenxing
    Niu, Hanlin
    Mahyuddin, Muhammad Nasiruddin
    Herrmann, Guido
    Carrasco, Joaquin
    [J]. 2021 21ST INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND SYSTEMS (ICCAS 2021), 2021, : 512 - 517
  • [6] Learning Representations in Model-Free Hierarchical Reinforcement Learning
    Rafati, Jacob
    Noelle, David C.
    [J]. THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 10009 - 10010
  • [7] Control of a Wave Energy Converter Using Model-free Deep Reinforcement Learning
    Chen, Kemeng
    Huang, Xuanrui
    Lin, Zechuan
    Xiao, Xi
    [J]. 2024 UKACC 14TH INTERNATIONAL CONFERENCE ON CONTROL, CONTROL, 2024, : 1 - 6
  • [8] A model-free deep reinforcement learning approach for control of exoskeleton gait patterns
    Rose, Lowell
    Bazzocchi, Michael C. F.
    Nejat, Goldie
    [J]. ROBOTICA, 2022, 40 (07) : 2189 - 2214
  • [9] Model-Free Quantum Gate Design and Calibration Using Deep Reinforcement Learning
    University of New South Wales, School of Engineering and Information Technology, Canberra
    ACT
    2600, Australia
    不详
    ACT
    2600, Australia
    不详
    QLD
    4111, Australia
    [J]. IEEE. Trans. Artif. Intell, 2024, 1 (346-357):
  • [10] Online Nonstochastic Model-Free Reinforcement Learning
    Ghai, Udaya
    Gupta, Arushi
    Xia, Wenhan
    Singh, Karan
    Hazan, Elad
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,